[ 636.185223] env[70013]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=70013) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 636.185604] env[70013]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=70013) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 636.185650] env[70013]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=70013) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 636.185984] env[70013]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 636.283281] env[70013]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=70013) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 636.293372] env[70013]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=70013) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 636.339186] env[70013]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 636.902920] env[70013]: INFO nova.virt.driver [None req-799e3e3d-819f-42ac-a636-a6613abfc44b None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 636.975071] env[70013]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.975241] env[70013]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.975355] env[70013]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=70013) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 639.914295] env[70013]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-47af1526-7a2f-4c5e-99f7-2c0957b97e68 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.930880] env[70013]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=70013) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 639.931146] env[70013]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-294c630a-f3df-444b-8600-4dadd90d312f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.966714] env[70013]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 83fb0. [ 639.967030] env[70013]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 2.992s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.967469] env[70013]: INFO nova.virt.vmwareapi.driver [None req-799e3e3d-819f-42ac-a636-a6613abfc44b None None] VMware vCenter version: 7.0.3 [ 639.971015] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f34d989-1d0d-4806-aee4-b509f0e2d72b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.989250] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddd28b0-4959-4333-8e33-f70588c81695 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.995611] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e693bf3-99bb-49fe-b26d-a7bd287e8ebd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.003365] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3f0d85-7dbe-49bd-a714-d786aede5dd7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.017227] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd0d2ae-c4c7-4cee-a269-f83358dd8afc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.023647] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea06d3c-7fea-4972-9ef4-aa9050490085 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.054721] env[70013]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-aa609e52-1923-4356-99d0-9467d377c6f4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.060740] env[70013]: DEBUG nova.virt.vmwareapi.driver [None req-799e3e3d-819f-42ac-a636-a6613abfc44b None None] Extension org.openstack.compute already exists. {{(pid=70013) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 640.063479] env[70013]: INFO nova.compute.provider_config [None req-799e3e3d-819f-42ac-a636-a6613abfc44b None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 640.567293] env[70013]: DEBUG nova.context [None req-799e3e3d-819f-42ac-a636-a6613abfc44b None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),9b1db4e2-c1f4-4367-af72-588a854be99e(cell1) {{(pid=70013) load_cells /opt/stack/nova/nova/context.py:472}} [ 640.567691] env[70013]: INFO nova.utils [None req-799e3e3d-819f-42ac-a636-a6613abfc44b None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 640.570048] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.570316] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.571040] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.571509] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Acquiring lock "9b1db4e2-c1f4-4367-af72-588a854be99e" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.571707] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Lock "9b1db4e2-c1f4-4367-af72-588a854be99e" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.572763] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Lock "9b1db4e2-c1f4-4367-af72-588a854be99e" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.593694] env[70013]: INFO dbcounter [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Registered counter for database nova_cell0 [ 640.602798] env[70013]: INFO dbcounter [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Registered counter for database nova_cell1 [ 640.606551] env[70013]: DEBUG oslo_db.sqlalchemy.engines [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=70013) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 640.607228] env[70013]: DEBUG oslo_db.sqlalchemy.engines [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=70013) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 640.612130] env[70013]: ERROR nova.db.main.api [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 640.612130] env[70013]: func(*args, **kwargs) [ 640.612130] env[70013]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 640.612130] env[70013]: self.work.run() [ 640.612130] env[70013]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 640.612130] env[70013]: result = self.fn(*self.args, **self.kwargs) [ 640.612130] env[70013]: File "/opt/stack/nova/nova/utils.py", line 697, in context_wrapper [ 640.612130] env[70013]: return func(*args, **kwargs) [ 640.612130] env[70013]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 640.612130] env[70013]: result = fn(*args, **kwargs) [ 640.612130] env[70013]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 640.612130] env[70013]: return f(*args, **kwargs) [ 640.612130] env[70013]: File "/opt/stack/nova/nova/objects/service.py", line 557, in _db_service_get_minimum_version [ 640.612130] env[70013]: return db.service_get_minimum_version(context, binaries) [ 640.612130] env[70013]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 640.612130] env[70013]: _check_db_access() [ 640.612130] env[70013]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 640.612130] env[70013]: stacktrace = ''.join(traceback.format_stack()) [ 640.612130] env[70013]: [ 640.613038] env[70013]: ERROR nova.db.main.api [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 640.613038] env[70013]: func(*args, **kwargs) [ 640.613038] env[70013]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 640.613038] env[70013]: self.work.run() [ 640.613038] env[70013]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 640.613038] env[70013]: result = self.fn(*self.args, **self.kwargs) [ 640.613038] env[70013]: File "/opt/stack/nova/nova/utils.py", line 697, in context_wrapper [ 640.613038] env[70013]: return func(*args, **kwargs) [ 640.613038] env[70013]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 640.613038] env[70013]: result = fn(*args, **kwargs) [ 640.613038] env[70013]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 640.613038] env[70013]: return f(*args, **kwargs) [ 640.613038] env[70013]: File "/opt/stack/nova/nova/objects/service.py", line 557, in _db_service_get_minimum_version [ 640.613038] env[70013]: return db.service_get_minimum_version(context, binaries) [ 640.613038] env[70013]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 640.613038] env[70013]: _check_db_access() [ 640.613038] env[70013]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 640.613038] env[70013]: stacktrace = ''.join(traceback.format_stack()) [ 640.613038] env[70013]: [ 640.613537] env[70013]: WARNING nova.objects.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 640.613669] env[70013]: WARNING nova.objects.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Failed to get minimum service version for cell 9b1db4e2-c1f4-4367-af72-588a854be99e [ 640.614129] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Acquiring lock "singleton_lock" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.614299] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Acquired lock "singleton_lock" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.614546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Releasing lock "singleton_lock" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.614878] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Full set of CONF: {{(pid=70013) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/service.py:357}} [ 640.615029] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ******************************************************************************** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 640.615160] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] Configuration options gathered from: {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 640.615294] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 640.615487] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 640.615739] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ================================================================================ {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 640.615975] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] allow_resize_to_same_host = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.616164] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] arq_binding_timeout = 300 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.616298] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] backdoor_port = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.616425] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] backdoor_socket = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.616589] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] block_device_allocate_retries = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.616752] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] block_device_allocate_retries_interval = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.616911] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cell_worker_thread_pool_size = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.617093] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cert = self.pem {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.617265] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.617434] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute_monitors = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.617623] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] config_dir = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.617800] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] config_drive_format = iso9660 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.617934] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.618109] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] config_source = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.618278] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] console_host = devstack {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.618442] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] control_exchange = nova {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.618609] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cpu_allocation_ratio = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.618794] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] daemon = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.618968] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] debug = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.619138] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] default_access_ip_network_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.619302] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] default_availability_zone = nova {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.619455] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] default_ephemeral_format = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.619613] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] default_green_pool_size = 1000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.619847] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.620020] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] default_schedule_zone = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.620187] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] disk_allocation_ratio = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.620344] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] enable_new_services = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.620519] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] enabled_apis = ['osapi_compute'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.620684] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] enabled_ssl_apis = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.620844] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] flat_injected = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.621009] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] force_config_drive = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.621177] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] force_raw_images = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.621344] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] graceful_shutdown_timeout = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.621503] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] heal_instance_info_cache_interval = -1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.621734] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] host = cpu-1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.621917] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.622158] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] initial_disk_allocation_ratio = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.622340] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] initial_ram_allocation_ratio = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.622570] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.622739] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instance_build_timeout = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.622902] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instance_delete_interval = 300 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.623089] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instance_format = [instance: %(uuid)s] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.623263] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instance_name_template = instance-%08x {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.623426] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instance_usage_audit = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.623597] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instance_usage_audit_period = month {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.623766] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.623930] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] instances_path = /opt/stack/data/nova/instances {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.624186] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] internal_service_availability_zone = internal {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.624354] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] key = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.624518] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] live_migration_retry_count = 30 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.624696] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_color = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.624860] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_config_append = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.625037] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.625205] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_dir = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.625365] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_file = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.625494] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_options = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.625766] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_rotate_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.625975] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_rotate_interval_type = days {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.626164] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] log_rotation_type = none {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.626297] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.626423] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.626593] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.626761] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.626890] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.627062] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] long_rpc_timeout = 1800 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.627227] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] max_concurrent_builds = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.627385] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] max_concurrent_live_migrations = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.627545] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] max_concurrent_snapshots = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.627729] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] max_local_block_devices = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.627897] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] max_logfile_count = 30 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.628072] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] max_logfile_size_mb = 200 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.628236] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] maximum_instance_delete_attempts = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.628404] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] metadata_listen = 0.0.0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.628577] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] metadata_listen_port = 8775 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.628771] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] metadata_workers = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.628938] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] migrate_max_retries = -1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.629119] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] mkisofs_cmd = genisoimage {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.629333] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] my_block_storage_ip = 10.180.1.21 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.629465] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] my_ip = 10.180.1.21 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.629672] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.629833] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] network_allocate_retries = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.630023] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.630192] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] osapi_compute_listen = 0.0.0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.630355] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] osapi_compute_listen_port = 8774 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.630524] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] osapi_compute_unique_server_name_scope = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.630692] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] osapi_compute_workers = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.630855] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] password_length = 12 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.631017] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] periodic_enable = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.631178] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] periodic_fuzzy_delay = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.631343] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] pointer_model = usbtablet {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.631506] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] preallocate_images = none {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.631663] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] publish_errors = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.631792] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] pybasedir = /opt/stack/nova {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.631946] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ram_allocation_ratio = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.632117] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] rate_limit_burst = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.632284] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] rate_limit_except_level = CRITICAL {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.632441] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] rate_limit_interval = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.632596] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] reboot_timeout = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.632753] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] reclaim_instance_interval = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.632908] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] record = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.633086] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] reimage_timeout_per_gb = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.633257] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] report_interval = 120 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.633418] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] rescue_timeout = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.633575] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] reserved_host_cpus = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.633733] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] reserved_host_disk_mb = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.633891] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] reserved_host_memory_mb = 512 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.634063] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] reserved_huge_pages = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.634230] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] resize_confirm_window = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.634391] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] resize_fs_using_block_device = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.634551] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] resume_guests_state_on_host_boot = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.634719] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.634895] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] rpc_response_timeout = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.635089] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] run_external_periodic_tasks = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.635267] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] running_deleted_instance_action = reap {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.635430] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] running_deleted_instance_poll_interval = 1800 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.635603] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] running_deleted_instance_timeout = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.635782] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler_instance_sync_interval = 120 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.636086] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_down_time = 720 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.636268] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] servicegroup_driver = db {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.636430] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] shell_completion = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.636592] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] shelved_offload_time = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.636755] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] shelved_poll_interval = 3600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.636921] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] shutdown_timeout = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.637094] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] source_is_ipv6 = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 640.637257] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ssl_only = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.091975] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.092576] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] sync_power_state_interval = 600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.092576] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] sync_power_state_pool_size = 1000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.092743] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] syslog_log_facility = LOG_USER {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.092828] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] tempdir = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.092909] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] timeout_nbd = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.093114] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] transport_url = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.093295] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] update_resources_interval = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.093461] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] use_cow_images = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.093626] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] use_journal = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.093790] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] use_json = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.093953] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] use_rootwrap_daemon = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.094131] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] use_stderr = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.094293] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] use_syslog = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.094451] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vcpu_pin_set = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.094624] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plugging_is_fatal = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.094795] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plugging_timeout = 300 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.094962] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] virt_mkfs = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.095141] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] volume_usage_poll_interval = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.095304] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] watch_log_file = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.095477] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] web = /usr/share/spice-html5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 641.095732] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.095975] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.096190] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.096373] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_concurrency.disable_process_locking = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.096703] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.096897] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.097080] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.097258] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.097430] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.097638] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.097808] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.auth_strategy = keystone {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.098022] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.compute_link_prefix = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.098220] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.098398] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.dhcp_domain = novalocal {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.098602] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.enable_instance_password = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.098734] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.glance_link_prefix = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.098900] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.099083] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.099251] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.instance_list_per_project_cells = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.099568] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.list_records_by_skipping_down_cells = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.099568] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.local_metadata_per_cell = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.099741] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.max_limit = 1000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.099903] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.metadata_cache_expiration = 15 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.100088] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.neutron_default_tenant_id = default {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.100270] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.response_validation = warn {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.100436] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.use_neutron_default_nets = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.100605] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.100765] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.100928] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.101112] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.101287] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.vendordata_dynamic_targets = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.101451] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.vendordata_jsonfile_path = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.101632] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.101865] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.backend = dogpile.cache.memcached {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.102060] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.backend_argument = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.102229] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.backend_expiration_time = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.102400] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.config_prefix = cache.oslo {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.102574] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.dead_timeout = 60.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.102740] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.debug_cache_backend = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.102905] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.enable_retry_client = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.103079] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.enable_socket_keepalive = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.103254] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.enabled = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.103421] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.enforce_fips_mode = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.103588] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.expiration_time = 600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.103756] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.hashclient_retry_attempts = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.103929] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.hashclient_retry_delay = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.104116] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_dead_retry = 300 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.104283] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_password = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.104448] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.104614] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.104786] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_pool_maxsize = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.104941] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.105113] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_sasl_enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.105296] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.105468] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_socket_timeout = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.105650] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.memcache_username = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.105839] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.proxies = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.106012] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.redis_db = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.106181] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.redis_password = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.106354] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.redis_sentinel_service_name = mymaster {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.106530] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.106766] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.redis_server = localhost:6379 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.106869] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.redis_socket_timeout = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.107051] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.redis_username = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.107222] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.retry_attempts = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.107390] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.retry_delay = 0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.107553] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.socket_keepalive_count = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.107742] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.socket_keepalive_idle = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.107910] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.socket_keepalive_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.108082] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.tls_allowed_ciphers = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.108248] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.tls_cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.108409] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.tls_certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.108573] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.tls_enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.108735] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cache.tls_keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.108907] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.109096] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.auth_type = password {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.109263] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.109443] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.catalog_info = volumev3::publicURL {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.109606] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.109773] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.109936] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.cross_az_attach = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.110112] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.debug = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.110277] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.endpoint_template = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.110441] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.http_retries = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.110605] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.110770] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.110945] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.os_region_name = RegionOne {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.111124] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.111289] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cinder.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.111463] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.111624] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.cpu_dedicated_set = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.111784] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.cpu_shared_set = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.111950] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.image_type_exclude_list = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.112126] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.112292] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.max_concurrent_disk_ops = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.112455] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.max_disk_devices_to_attach = -1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.112617] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.112789] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.112954] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.resource_provider_association_refresh = 300 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.113128] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.113293] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.shutdown_retry_interval = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.113475] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.113656] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] conductor.workers = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.113867] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] console.allowed_origins = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.114063] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] console.ssl_ciphers = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.114245] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] console.ssl_minimum_version = default {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.114419] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] consoleauth.enforce_session_timeout = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.114590] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] consoleauth.token_ttl = 600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.114761] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.114921] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.115098] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.115264] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.115425] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.115589] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.115752] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.115912] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.116091] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.116257] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.116415] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.116574] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.116734] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.116904] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.service_type = accelerator {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.117076] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.117239] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.117398] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.117557] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.117768] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.117939] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] cyborg.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.118134] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.asyncio_connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.118288] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.asyncio_slave_connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.118461] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.backend = sqlalchemy {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.118657] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.118832] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.connection_debug = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.119023] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.connection_parameters = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.119187] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.connection_recycle_time = 3600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.119352] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.connection_trace = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.119515] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.db_inc_retry_interval = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.119678] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.db_max_retries = 20 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.119841] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.db_max_retry_interval = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.120010] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.db_retry_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.120188] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.max_overflow = 50 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.120350] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.max_pool_size = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.120511] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.max_retries = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.120680] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.120839] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.mysql_wsrep_sync_wait = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.120999] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.pool_timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.121176] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.retry_interval = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.121336] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.slave_connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.121498] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.sqlite_synchronous = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.121661] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] database.use_db_reconnect = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.121830] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.asyncio_connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.121988] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.asyncio_slave_connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.122171] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.backend = sqlalchemy {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.122340] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.122505] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.connection_debug = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.122677] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.connection_parameters = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.122841] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.connection_recycle_time = 3600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.123015] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.connection_trace = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.123183] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.db_inc_retry_interval = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.123346] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.db_max_retries = 20 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.123509] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.db_max_retry_interval = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.123672] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.db_retry_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.123837] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.max_overflow = 50 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.124006] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.max_pool_size = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.124177] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.max_retries = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.124347] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.124507] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.124667] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.pool_timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.124830] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.retry_interval = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.124989] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.slave_connection = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.125164] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] api_database.sqlite_synchronous = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.125341] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] devices.enabled_mdev_types = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.125518] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.125725] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ephemeral_storage_encryption.default_format = luks {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.125904] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ephemeral_storage_encryption.enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.126080] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.126257] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.api_servers = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.126425] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.126590] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.126756] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.126915] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.127089] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.127256] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.debug = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.127423] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.default_trusted_certificate_ids = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.127596] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.enable_certificate_validation = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.127784] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.enable_rbd_download = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.127949] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.128223] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.128394] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.128561] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.128764] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.128936] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.num_retries = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.129122] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.rbd_ceph_conf = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.129289] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.rbd_connect_timeout = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.129464] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.rbd_pool = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.129643] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.rbd_user = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.129807] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.129968] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.130141] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.130311] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.service_type = image {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.130474] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.130632] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.130792] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.130951] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.131145] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.131312] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.verify_glance_signatures = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.131473] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] glance.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.131649] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] guestfs.debug = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.131889] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.132085] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.auth_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.132253] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.132417] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.132583] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.132746] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.132906] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.133080] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.133250] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.133409] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.133570] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.133727] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.133886] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.134055] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.134217] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.134386] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.service_type = shared-file-system {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.134549] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.share_apply_policy_timeout = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.134713] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.134873] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.135041] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.135206] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.135389] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.135577] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] manila.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.135774] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] mks.enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.136161] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.136356] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] image_cache.manager_interval = 2400 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.136524] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] image_cache.precache_concurrency = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.136694] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] image_cache.remove_unused_base_images = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.136865] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.137043] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.137225] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] image_cache.subdirectory_name = _base {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.137401] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.api_max_retries = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.137575] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.api_retry_interval = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.137785] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.137960] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.auth_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.138138] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.138301] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.138463] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.138646] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.conductor_group = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.138821] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.138982] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.139158] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.139323] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.139482] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.139643] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.139802] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.139967] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.peer_list = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.140139] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.140304] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.140464] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.serial_console_state_timeout = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.140623] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.140793] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.service_type = baremetal {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.140953] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.shard = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.141129] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.141290] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.141448] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.141607] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.141789] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.141953] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ironic.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.142150] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.142329] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] key_manager.fixed_key = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.142514] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.142680] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.barbican_api_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.142843] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.barbican_endpoint = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.143023] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.barbican_endpoint_type = public {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.143191] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.barbican_region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.143356] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.143516] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.143680] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.143896] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.144103] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.144275] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.number_of_retries = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.144437] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.retry_delay = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.144599] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.send_service_user_token = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.144764] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.144923] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.145097] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.verify_ssl = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.145261] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican.verify_ssl_path = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.145429] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.145631] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.auth_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.145802] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.145961] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.146139] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.146303] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.146462] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.146623] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.146782] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] barbican_service_user.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.146949] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.approle_role_id = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.147121] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.approle_secret_id = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.147293] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.kv_mountpoint = secret {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.147452] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.kv_path = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.147633] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.kv_version = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.147804] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.namespace = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.147966] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.root_token_id = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.148141] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.ssl_ca_crt_file = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.148310] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.timeout = 60.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.148471] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.use_ssl = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.148659] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.148843] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.149013] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.149188] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.149352] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.149513] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.149702] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.149897] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.150071] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.150245] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.150409] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.150567] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.150727] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.150883] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.151067] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.service_type = identity {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.151234] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.151394] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.151553] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.151709] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.151887] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.152058] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] keystone.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.152252] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.ceph_mount_options = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.152575] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.152760] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.connection_uri = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.152925] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.cpu_mode = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.153108] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.cpu_model_extra_flags = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.153284] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.cpu_models = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.153457] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.cpu_power_governor_high = performance {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.153627] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.cpu_power_governor_low = powersave {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.153800] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.cpu_power_management = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.153977] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.154164] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.device_detach_attempts = 8 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.154331] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.device_detach_timeout = 20 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.154500] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.disk_cachemodes = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.154663] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.disk_prefix = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.154831] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.enabled_perf_events = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.154997] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.file_backed_memory = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.155180] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.gid_maps = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.155341] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.hw_disk_discard = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.155515] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.hw_machine_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.155701] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.images_rbd_ceph_conf = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.155871] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.156045] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.156220] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.images_rbd_glance_store_name = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.156388] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.images_rbd_pool = rbd {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.156558] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.images_type = default {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.156722] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.images_volume_group = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.156887] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.inject_key = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.157064] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.inject_partition = -2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.157244] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.inject_password = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.157472] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.iscsi_iface = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.157677] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.iser_use_multipath = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.157856] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_bandwidth = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.158036] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.158209] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_downtime = 500 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.158375] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.158539] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.158704] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_inbound_addr = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.158867] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.159164] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_permit_post_copy = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.159337] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_scheme = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.159517] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_timeout_action = abort {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.159686] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_tunnelled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.159852] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_uri = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.160057] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.live_migration_with_native_tls = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.160197] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.max_queues = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.160366] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.160598] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.160768] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.nfs_mount_options = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.161095] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.161278] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.161450] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.num_iser_scan_tries = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.161615] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.num_memory_encrypted_guests = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.161813] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.161996] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.num_pcie_ports = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.162187] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.num_volume_scan_tries = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.162364] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.pmem_namespaces = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.162531] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.quobyte_client_cfg = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.162837] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.163030] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rbd_connect_timeout = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.163207] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.163381] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.163535] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rbd_secret_uuid = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.163697] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rbd_user = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.163871] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.164053] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.remote_filesystem_transport = ssh {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.164221] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rescue_image_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.164383] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rescue_kernel_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.164541] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rescue_ramdisk_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.164709] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.164872] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.rx_queue_size = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.165051] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.smbfs_mount_options = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.165362] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.165554] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.snapshot_compression = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.165729] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.snapshot_image_format = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.165964] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.166149] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.sparse_logical_volumes = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.166315] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.swtpm_enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.166484] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.swtpm_group = tss {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.166651] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.swtpm_user = tss {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.166820] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.sysinfo_serial = unique {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.166979] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.tb_cache_size = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.167158] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.tx_queue_size = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.167326] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.uid_maps = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.167488] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.use_virtio_for_bridges = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.167692] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.virt_type = kvm {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.167872] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.volume_clear = zero {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.168050] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.volume_clear_size = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.168225] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.volume_enforce_multipath = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.168395] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.volume_use_multipath = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.168564] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.vzstorage_cache_path = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.168753] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.168929] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.vzstorage_mount_group = qemu {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.169108] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.vzstorage_mount_opts = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.169283] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.169589] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.169781] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.vzstorage_mount_user = stack {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.169954] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.170151] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.170334] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.auth_type = password {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.170498] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.170663] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.170833] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.170996] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.171177] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.171353] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.default_floating_pool = public {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.171519] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.171685] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.extension_sync_interval = 600 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.171854] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.http_retries = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.172029] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.172198] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.172362] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.172535] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.172720] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.172911] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.ovs_bridge = br-int {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.173100] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.physnets = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.173278] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.region_name = RegionOne {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.173441] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.173612] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.service_metadata_proxy = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.173812] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.174020] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.service_type = network {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.174198] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.174363] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.174526] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.174686] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.174872] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.175049] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] neutron.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.175228] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] notifications.bdms_in_notifications = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.175408] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] notifications.default_level = INFO {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.175590] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] notifications.include_share_mapping = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.175769] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] notifications.notification_format = unversioned {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.175935] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] notifications.notify_on_state_change = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.176125] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.176310] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] pci.alias = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.176482] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] pci.device_spec = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.176648] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] pci.report_in_placement = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.176822] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.176994] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.auth_type = password {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.177179] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.177341] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.177500] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.177700] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.177882] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.178056] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.178224] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.default_domain_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.178386] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.default_domain_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.178545] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.domain_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.178734] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.domain_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.178901] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.179078] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.179241] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.179402] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.179558] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.179726] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.password = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.179890] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.project_domain_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.180069] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.project_domain_name = Default {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.180246] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.project_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.180418] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.project_name = service {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.180590] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.region_name = RegionOne {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.180816] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.181028] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.181213] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.service_type = placement {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.181379] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.181541] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.181702] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.181866] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.system_scope = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.182033] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.182200] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.trust_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.182360] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.user_domain_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.182527] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.user_domain_name = Default {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.182688] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.user_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.182864] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.username = nova {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.183064] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.183233] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] placement.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.183412] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.cores = 20 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.183578] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.count_usage_from_placement = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.183806] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.183990] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.injected_file_content_bytes = 10240 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.184178] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.injected_file_path_length = 255 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.184347] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.injected_files = 5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.184513] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.instances = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.184680] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.key_pairs = 100 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.184847] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.metadata_items = 128 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.185021] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.ram = 51200 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.185195] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.recheck_quota = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.185366] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.server_group_members = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.185569] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.server_groups = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.185805] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.185987] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] quota.unified_limits_resource_strategy = require {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.186178] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.186345] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.186506] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.image_metadata_prefilter = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.186669] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.186860] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.max_attempts = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.187046] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.max_placement_results = 1000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.187218] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.187382] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.query_placement_for_image_type_support = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.187544] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.187741] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] scheduler.workers = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.187928] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.188116] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.188301] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.188479] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.188647] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.188813] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.188978] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.189185] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.189355] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.host_subset_size = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.189525] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.189686] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.189887] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.190116] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.190295] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.190467] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.isolated_hosts = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.190658] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.isolated_images = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.190842] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.191025] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.191190] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.191356] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.pci_in_placement = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.191523] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.191688] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.191856] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.192029] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.192200] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.192367] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.192527] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.track_instance_changes = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.192760] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.192971] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] metrics.required = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.193162] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] metrics.weight_multiplier = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.193333] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.193504] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] metrics.weight_setting = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.193832] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.194021] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] serial_console.enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.194210] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] serial_console.port_range = 10000:20000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.194387] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.194563] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.194735] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] serial_console.serialproxy_port = 6083 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.194909] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.195098] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.auth_type = password {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.195269] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.195432] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.195597] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.195761] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.195943] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.196140] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.send_service_user_token = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.196312] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.196475] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] service_user.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.196673] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.agent_enabled = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.196852] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.197181] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.197395] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.197569] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.html5proxy_port = 6082 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.197759] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.image_compression = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.197928] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.jpeg_compression = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.198102] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.playback_compression = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.198272] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.require_secure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.198444] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.server_listen = 127.0.0.1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.198637] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.198950] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.199142] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.streaming_mode = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.199311] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] spice.zlib_compression = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.199481] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] upgrade_levels.baseapi = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.199659] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] upgrade_levels.compute = auto {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.199825] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] upgrade_levels.conductor = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.200034] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] upgrade_levels.scheduler = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.200218] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.200385] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.auth_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.200545] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.200704] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.200867] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.201038] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.201205] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.201366] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.201524] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vendordata_dynamic_auth.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.201734] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.api_retry_count = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.201895] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.ca_file = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.202085] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.cache_prefix = devstack-image-cache {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.202261] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.cluster_name = testcl1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.202424] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.connection_pool_size = 10 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.202585] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.console_delay_seconds = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.202756] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.datastore_regex = ^datastore.* {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.202982] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.203180] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.host_password = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.203350] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.host_port = 443 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.203522] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.host_username = administrator@vsphere.local {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.203693] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.insecure = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.203859] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.integration_bridge = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.204034] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.maximum_objects = 100 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.204202] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.pbm_default_policy = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.204365] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.pbm_enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.204525] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.pbm_wsdl_location = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.204729] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.204914] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.serial_port_proxy_uri = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.205088] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.serial_port_service_uri = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.205266] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.task_poll_interval = 0.5 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.205439] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.use_linked_clone = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.205614] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.vnc_keymap = en-us {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.205794] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.vnc_port = 5900 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.205971] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vmware.vnc_port_total = 10000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.206176] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.auth_schemes = ['none'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.206356] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.206660] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.206850] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.207029] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.novncproxy_port = 6080 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.207227] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.server_listen = 127.0.0.1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.207407] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.207594] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.vencrypt_ca_certs = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.207782] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.vencrypt_client_cert = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.207958] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vnc.vencrypt_client_key = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.208152] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.208319] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.disable_deep_image_inspection = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.208484] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.208671] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.208851] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.209026] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.disable_rootwrap = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.209197] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.enable_numa_live_migration = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.209360] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.209522] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.209683] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.209844] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.libvirt_disable_apic = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.210009] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.210184] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.210345] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.210506] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.210668] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.210909] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.211126] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.211297] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.211460] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.211627] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.211815] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.211982] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] wsgi.secure_proxy_ssl_header = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.212165] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] zvm.ca_file = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.212329] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] zvm.cloud_connector_url = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.212809] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.212997] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] zvm.reachable_timeout = 300 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.213192] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.213384] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.213574] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.connection_string = messaging:// {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.213760] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.enabled = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.213958] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.es_doc_type = notification {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.214145] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.es_scroll_size = 10000 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.214320] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.es_scroll_time = 2m {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.214487] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.filter_error_trace = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.214661] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.hmac_keys = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.214834] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.sentinel_service_name = mymaster {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.215012] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.socket_timeout = 0.1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.215184] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.trace_requests = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.215349] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler.trace_sqlalchemy = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.215533] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler_jaeger.process_tags = {} {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.215696] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler_jaeger.service_name_prefix = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.215860] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] profiler_otlp.service_name_prefix = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.216047] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.216217] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.216379] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.216539] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.216727] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.216929] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.217111] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.217277] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.217440] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.217635] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.217814] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.217989] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.218175] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.218343] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.218509] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.218706] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.218878] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.219055] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.219229] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.219393] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.219556] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.219717] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.219895] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.220104] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.220278] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.220448] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.220610] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.220776] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.220939] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.221114] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.221280] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.ssl = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.221453] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.221619] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.221781] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.221947] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.222153] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.ssl_version = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.222336] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.222524] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.222694] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_notifications.retry = -1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.222873] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.223088] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_messaging_notifications.transport_url = **** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.223278] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.auth_section = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.223449] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.auth_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.223613] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.cafile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.223777] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.certfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.223943] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.collect_timing = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.224119] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.connect_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.224282] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.connect_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.224443] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.endpoint_id = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.224616] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.endpoint_interface = publicURL {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.224780] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.endpoint_override = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.224941] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.endpoint_region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.225113] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.endpoint_service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.225278] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.endpoint_service_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.225440] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.insecure = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.225600] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.keyfile = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.225762] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.max_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.225921] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.min_version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.226137] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.region_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.226314] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.retriable_status_codes = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.226476] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.service_name = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.226637] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.service_type = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.226804] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.split_loggers = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.226964] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.status_code_retries = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.227146] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.status_code_retry_delay = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.227305] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.timeout = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.227466] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.valid_interfaces = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.227655] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_limit.version = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.227833] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_reports.file_event_handler = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.228009] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.228179] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] oslo_reports.log_dir = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.228354] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.228516] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.228722] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.228875] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.229069] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.229260] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.229426] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.229599] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.229761] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_ovs_privileged.group = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.229921] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.230096] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.230268] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.230434] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.230593] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] vif_plug_ovs_privileged.user = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.230765] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.flat_interface = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.230945] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.231140] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.231310] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.231482] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.231656] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.231825] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.231989] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.232208] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.232391] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_ovs.isolate_vif = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.232569] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.232735] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.232907] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.233090] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_ovs.ovsdb_interface = native {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.233258] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] os_vif_ovs.per_port_bridge = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.233429] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] privsep_osbrick.capabilities = [21] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.233587] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] privsep_osbrick.group = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.233747] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] privsep_osbrick.helper_command = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.233907] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] privsep_osbrick.log_daemon_traceback = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.234081] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.234249] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.234408] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] privsep_osbrick.user = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.234580] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.234739] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] nova_sys_admin.group = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.234894] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] nova_sys_admin.helper_command = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.235066] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] nova_sys_admin.log_daemon_traceback = False {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.235262] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.235428] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.235586] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] nova_sys_admin.user = None {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 641.235721] env[70013]: DEBUG oslo_service.backend.eventlet.service [None req-f00723b9-8923-4404-80b6-423c7b045cb4 None None] ******************************************************************************** {{(pid=70013) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 641.236154] env[70013]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 641.739368] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Getting list of instances from cluster (obj){ [ 641.739368] env[70013]: value = "domain-c8" [ 641.739368] env[70013]: _type = "ClusterComputeResource" [ 641.739368] env[70013]: } {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 641.740728] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a97a10-d35a-46d2-9da1-67a47977970c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.749793] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Got total of 0 instances {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 641.750458] env[70013]: WARNING nova.virt.vmwareapi.driver [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 641.750918] env[70013]: INFO nova.virt.node [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Generated node identity dc0a4c9d-b611-453e-a900-41e280c29b95 [ 641.751165] env[70013]: INFO nova.virt.node [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Wrote node identity dc0a4c9d-b611-453e-a900-41e280c29b95 to /opt/stack/data/n-cpu-1/compute_id [ 642.254311] env[70013]: WARNING nova.compute.manager [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Compute nodes ['dc0a4c9d-b611-453e-a900-41e280c29b95'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 643.261862] env[70013]: INFO nova.compute.manager [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 644.270072] env[70013]: WARNING nova.compute.manager [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 644.270072] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.270072] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.270072] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.270072] env[70013]: DEBUG nova.compute.resource_tracker [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 644.270072] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b39320ef-2b43-40d0-8a5d-39b9cd1f1ab3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.278585] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4775012-3b59-47ba-9935-390603afecb2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.294826] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8c7eaf-a685-4c4e-ab64-2f2f981df690 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.302385] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c53a2b-5282-4194-9a9d-2a07160f423e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.334435] env[70013]: DEBUG nova.compute.resource_tracker [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180794MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 644.334998] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.335470] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.840299] env[70013]: WARNING nova.compute.resource_tracker [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] No compute node record for cpu-1:dc0a4c9d-b611-453e-a900-41e280c29b95: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host dc0a4c9d-b611-453e-a900-41e280c29b95 could not be found. [ 645.343695] env[70013]: INFO nova.compute.resource_tracker [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: dc0a4c9d-b611-453e-a900-41e280c29b95 [ 646.854027] env[70013]: DEBUG nova.compute.resource_tracker [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 646.854027] env[70013]: DEBUG nova.compute.resource_tracker [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 647.008567] env[70013]: INFO nova.scheduler.client.report [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] [req-42fdcf56-2271-4285-9857-8b545db95673] Created resource provider record via placement API for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 647.028024] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5f693f-c3cb-4ad3-a6da-611a88943caf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.035312] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3e6dcc-774b-4052-bcf9-649b9e2c491e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.067070] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f76ec8-b299-41a5-abb8-4c7ce17395e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.075459] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374d23ac-ad4f-4c1b-ab8b-ff184892fa8f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.089343] env[70013]: DEBUG nova.compute.provider_tree [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 647.626113] env[70013]: DEBUG nova.scheduler.client.report [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Updated inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 647.626503] env[70013]: DEBUG nova.compute.provider_tree [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Updating resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 generation from 0 to 1 during operation: update_inventory {{(pid=70013) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 647.626782] env[70013]: DEBUG nova.compute.provider_tree [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 647.678027] env[70013]: DEBUG nova.compute.provider_tree [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Updating resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 generation from 1 to 2 during operation: update_traits {{(pid=70013) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 648.182576] env[70013]: DEBUG nova.compute.resource_tracker [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 648.182972] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.847s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.183391] env[70013]: DEBUG nova.service [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Creating RPC server for service compute {{(pid=70013) start /opt/stack/nova/nova/service.py:177}} [ 648.196934] env[70013]: DEBUG nova.service [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] Join ServiceGroup membership for this service compute {{(pid=70013) start /opt/stack/nova/nova/service.py:194}} [ 648.197242] env[70013]: DEBUG nova.servicegroup.drivers.db [None req-e6bc0d4e-1d79-4bd7-ad4d-da54115b6fe2 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=70013) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 686.139325] env[70013]: INFO nova.utils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] The default thread pool MainProcess.default is initialized [ 686.139707] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.139908] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.647705] env[70013]: DEBUG nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 687.185749] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.186094] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 687.188682] env[70013]: INFO nova.compute.claims [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.325718] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "98a0199a-4209-4faf-adf5-7ae33b099d20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.325718] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 687.828531] env[70013]: DEBUG nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 688.270402] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575e372d-c5c6-4d32-b1b0-cf1b62cb56ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.285243] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb17509a-250d-4ed3-9917-877225ba471c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.336422] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e5e226-f4a6-4c84-a9e9-4add5e8128eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.354305] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b344b6-f005-4348-8b32-d42cd495d4a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.377572] env[70013]: DEBUG nova.compute.provider_tree [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.382845] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 688.884446] env[70013]: DEBUG nova.scheduler.client.report [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.393850] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.394689] env[70013]: DEBUG nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 689.401262] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.020s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.401958] env[70013]: INFO nova.compute.claims [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.915595] env[70013]: DEBUG nova.compute.utils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 689.920646] env[70013]: DEBUG nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 689.920646] env[70013]: DEBUG nova.network.neutron [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 690.418710] env[70013]: DEBUG nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 690.506234] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4d3ec4-f7aa-425b-afdd-fc48b80f8c10 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.516202] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7250fd-3756-4acc-b36c-1074520fc46b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.555112] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f261ff9-f3fd-45e5-9b76-331211621e9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.566963] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e31007-8d7c-426f-871c-b87f0ec08710 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.583835] env[70013]: DEBUG nova.compute.provider_tree [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.619730] env[70013]: DEBUG nova.policy [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecdb0c78f89f4908a1f7f21f4c82b771', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7691ed6dbc544b6e892c72a4235deac8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 691.089344] env[70013]: DEBUG nova.scheduler.client.report [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.435706] env[70013]: DEBUG nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 691.469820] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 691.470724] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.471076] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 691.471575] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.471859] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 691.472147] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 691.472488] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 691.472758] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 691.473326] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 691.473619] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 691.473908] env[70013]: DEBUG nova.virt.hardware [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 691.475807] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cdbdea-9a59-4056-aa64-b7fe5e3b1820 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.488247] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ea5ca9-9781-41c8-a0f8-67557cbca08b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.505808] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8409638b-8861-4ae0-a9b7-1c5bf0173f4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.604908] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.201s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.604908] env[70013]: DEBUG nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 692.108509] env[70013]: DEBUG nova.compute.utils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 692.110099] env[70013]: DEBUG nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 692.265596] env[70013]: DEBUG nova.network.neutron [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Successfully created port: ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.612490] env[70013]: DEBUG nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 692.900027] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.901364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.403609] env[70013]: DEBUG nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 693.457553] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.457925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.627425] env[70013]: DEBUG nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 693.653705] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 693.653956] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.654184] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 693.654320] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.654465] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 693.654613] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 693.654847] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 693.656534] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 693.656534] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 693.656609] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 693.656832] env[70013]: DEBUG nova.virt.hardware [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 693.657673] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c9e110-6bf8-47ad-af1e-3e21fe26d66c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.669815] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8cc738-fd9c-453c-9e63-7e6d83c3efd7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.687083] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 693.699074] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.699666] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9bf149d-3725-499f-b25e-3fed92c0da35 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.712447] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Created folder: OpenStack in parent group-v4. [ 693.712847] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Creating folder: Project (0497ea2ac34845cc9030211b2fd3c803). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.713245] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90a22a1b-945c-4fca-b38a-56522058b01e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.724420] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Created folder: Project (0497ea2ac34845cc9030211b2fd3c803) in parent group-v836999. [ 693.724420] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Creating folder: Instances. Parent ref: group-v837000. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.724420] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-340c5832-7a7e-4eb8-bd7d-6fa422553a7d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.735104] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Created folder: Instances in parent group-v837000. [ 693.736077] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 693.736077] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 693.736077] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf7fc77b-023d-4bc3-aaa9-bde9eb8a32bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.753485] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 693.753485] env[70013]: value = "task-4230460" [ 693.753485] env[70013]: _type = "Task" [ 693.753485] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.763509] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230460, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.937934] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.938263] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.939968] env[70013]: INFO nova.compute.claims [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.961406] env[70013]: DEBUG nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 694.270116] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230460, 'name': CreateVM_Task, 'duration_secs': 0.314612} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.270248] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 694.271455] env[70013]: DEBUG oslo_vmware.service [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332dc74f-00d6-4ab0-ad10-8fc5a57010d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.282797] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.282797] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 694.283491] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 694.283780] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c096a6cf-ccc6-43bb-a25d-da8ed48f14ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.290647] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 694.290647] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf4388-c011-8192-590f-7767a1915d96" [ 694.290647] env[70013]: _type = "Task" [ 694.290647] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.309766] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 694.309880] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 694.310499] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.310499] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 694.310831] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 694.314306] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9562f967-1999-4b1a-b404-f50ac87fa1f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.317588] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "417f8c74-81cc-47d6-80ca-b61d03ebec50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 694.317772] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.336933] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 694.337106] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 694.338203] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd9c905-cffd-4389-ac5e-0310cc337974 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.347950] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c22b3f2-4e9d-4cec-8713-6eb73e8b09a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.357225] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 694.357225] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf1f61-28cb-8f40-35bc-86c43f394966" [ 694.357225] env[70013]: _type = "Task" [ 694.357225] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.366511] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf1f61-28cb-8f40-35bc-86c43f394966, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.497627] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 694.825806] env[70013]: DEBUG nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 694.870970] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Preparing fetch location {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 694.871440] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Creating directory with path [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 694.871975] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69fab4be-b65a-47b4-8e8d-0a0b8fffb649 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.894738] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Created directory with path [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 694.894948] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Fetch image to [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 694.895183] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Downloading image file data 0ddeef65-a14c-4065-97bb-58b607968d19 to [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk on the data store datastore2 {{(pid=70013) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 694.897118] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab95f4eb-baa6-4991-a547-741ee7df7e33 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.907110] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82275e1-f70d-442c-b14e-ec8db0916904 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.922368] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08207012-a3df-4557-9fbf-689b6f0df584 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.973170] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0a4682-86ea-458a-80fa-be6cb52a8ea5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.980748] env[70013]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-93ed7eb2-e4f4-4144-87f7-05944ad73a01 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.015097] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Downloading image file data 0ddeef65-a14c-4065-97bb-58b607968d19 to the data store datastore2 {{(pid=70013) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 695.098301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "3b29c99a-0d55-40b1-a155-199b1ced2146" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.098570] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.122741] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d1738a-e5ad-4fb1-b7bb-f8dba812c88b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.133731] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a070a415-fbb6-4f82-ab67-5d07ceb55427 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.179094] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56c7eec-3871-4a24-972f-cdc53f265eb7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.192802] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76eab2c-e5dd-4a7f-8f38-1ea05c1da539 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.206234] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "9a553d85-5086-46c5-8df4-451928e38ed9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.207283] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "9a553d85-5086-46c5-8df4-451928e38ed9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.220149] env[70013]: DEBUG nova.compute.provider_tree [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.256341] env[70013]: DEBUG oslo_vmware.rw_handles [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=70013) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 695.355336] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.608495] env[70013]: DEBUG nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 695.725238] env[70013]: DEBUG nova.scheduler.client.report [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 695.732741] env[70013]: DEBUG nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 695.923557] env[70013]: DEBUG oslo_vmware.rw_handles [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Completed reading data from the image iterator. {{(pid=70013) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 695.923557] env[70013]: DEBUG oslo_vmware.rw_handles [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 696.074860] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Downloaded image file data 0ddeef65-a14c-4065-97bb-58b607968d19 to vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk on the data store datastore2 {{(pid=70013) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 696.076543] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Caching image {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 696.076793] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Copying Virtual Disk [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk to [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 696.077098] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-292f47ba-c091-416f-9b86-3a4b4c596613 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.090317] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 696.090317] env[70013]: value = "task-4230461" [ 696.090317] env[70013]: _type = "Task" [ 696.090317] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.101063] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230461, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.135297] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.245279] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.246224] env[70013]: DEBUG nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 696.251925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.754s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.254352] env[70013]: INFO nova.compute.claims [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.274273] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.590481] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "cf8be667-da84-466b-9b7b-3bdfda6a8193" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.590603] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "cf8be667-da84-466b-9b7b-3bdfda6a8193" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.612691] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230461, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.646016] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.646016] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.735104] env[70013]: DEBUG nova.network.neutron [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Successfully updated port: ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 696.756124] env[70013]: DEBUG nova.compute.utils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 696.761748] env[70013]: DEBUG nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 696.765821] env[70013]: DEBUG nova.network.neutron [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 697.030783] env[70013]: DEBUG nova.policy [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a7386d27973488ea728d9f442f3b47e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbdf4732a346424b8ff1c39c7e900f08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 697.104854] env[70013]: DEBUG nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 697.108548] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230461, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691618} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.108884] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Copied Virtual Disk [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk to [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 697.109089] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleting the datastore file [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 697.109382] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16504f7f-eb41-4fed-b359-e974bea86f52 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.118779] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 697.118779] env[70013]: value = "task-4230462" [ 697.118779] env[70013]: _type = "Task" [ 697.118779] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.130391] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230462, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.149995] env[70013]: DEBUG nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 697.198893] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.199384] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.199612] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.200252] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.200252] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.200252] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.200427] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_power_states {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.241225] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "refresh_cache-2b91ec52-2d32-4f86-acf2-f74f661f65b5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.241416] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquired lock "refresh_cache-2b91ec52-2d32-4f86-acf2-f74f661f65b5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.241860] env[70013]: DEBUG nova.network.neutron [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.275106] env[70013]: DEBUG nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 697.457513] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d838a4b1-c2fe-4a2e-b082-84488f2d8b4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.466791] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2b881b-3b6d-4140-ac0a-db18d7ae212d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.498395] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d20a0ef-47ca-4520-9249-3741bf8f062f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.508722] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adbd8af3-df67-45c0-a097-98526ffff78c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.523776] env[70013]: DEBUG nova.compute.provider_tree [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.642224] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230462, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026208} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.642224] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 697.642224] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Moving file from [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5/0ddeef65-a14c-4065-97bb-58b607968d19 to [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19. {{(pid=70013) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 697.642224] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-8d2d4039-8613-42e7-a689-fbbc93c65b6c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.646099] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.658660] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 697.658660] env[70013]: value = "task-4230463" [ 697.658660] env[70013]: _type = "Task" [ 697.658660] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.676212] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230463, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.683285] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.705035] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Getting list of instances from cluster (obj){ [ 697.705035] env[70013]: value = "domain-c8" [ 697.705035] env[70013]: _type = "ClusterComputeResource" [ 697.705035] env[70013]: } {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 697.706106] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db1f17d-69a4-4ab5-95a0-c7f914871a8f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.720464] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Got total of 1 instances {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 697.720464] env[70013]: WARNING nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] While synchronizing instance power states, found 4 instances in the database and 1 instances on the hypervisor. [ 697.720464] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 2b91ec52-2d32-4f86-acf2-f74f661f65b5 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 697.720464] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 98a0199a-4209-4faf-adf5-7ae33b099d20 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 697.720464] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid e8998cab-533b-46b1-a1ea-5eb5db7d6d99 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 697.720829] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 9c5f2422-820e-454b-98be-695e3ac6d1e3 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 697.721024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.722301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "98a0199a-4209-4faf-adf5-7ae33b099d20" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.722301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.722301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.722301] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.722301] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 697.722506] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.888018] env[70013]: DEBUG nova.network.neutron [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.026825] env[70013]: DEBUG nova.scheduler.client.report [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 698.037578] env[70013]: DEBUG nova.compute.manager [req-e6b4f5e3-d1e1-4638-b8e1-8999aa458a70 req-58a26502-46ad-429f-acc3-786dd2ccddb8 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Received event network-vif-plugged-ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 698.040162] env[70013]: DEBUG oslo_concurrency.lockutils [req-e6b4f5e3-d1e1-4638-b8e1-8999aa458a70 req-58a26502-46ad-429f-acc3-786dd2ccddb8 service nova] Acquiring lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.040162] env[70013]: DEBUG oslo_concurrency.lockutils [req-e6b4f5e3-d1e1-4638-b8e1-8999aa458a70 req-58a26502-46ad-429f-acc3-786dd2ccddb8 service nova] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.040162] env[70013]: DEBUG oslo_concurrency.lockutils [req-e6b4f5e3-d1e1-4638-b8e1-8999aa458a70 req-58a26502-46ad-429f-acc3-786dd2ccddb8 service nova] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.040162] env[70013]: DEBUG nova.compute.manager [req-e6b4f5e3-d1e1-4638-b8e1-8999aa458a70 req-58a26502-46ad-429f-acc3-786dd2ccddb8 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] No waiting events found dispatching network-vif-plugged-ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 698.040162] env[70013]: WARNING nova.compute.manager [req-e6b4f5e3-d1e1-4638-b8e1-8999aa458a70 req-58a26502-46ad-429f-acc3-786dd2ccddb8 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Received unexpected event network-vif-plugged-ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 for instance with vm_state building and task_state spawning. [ 698.179647] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230463, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.02684} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.179890] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] File moved {{(pid=70013) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 698.180072] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Cleaning up location [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 698.180237] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleting the datastore file [datastore2] vmware_temp/9c8055e9-1f4f-4308-a9e1-f11d7a2a3cf5 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 698.180494] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d370d9e5-a83d-496f-8a82-5112e92b6a20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.189589] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 698.189589] env[70013]: value = "task-4230464" [ 698.189589] env[70013]: _type = "Task" [ 698.189589] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.200859] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.230334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.293044] env[70013]: DEBUG nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 698.331468] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 698.331468] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.331468] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 698.331697] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.332141] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 698.332141] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 698.332922] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 698.333075] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 698.333282] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 698.333566] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 698.333617] env[70013]: DEBUG nova.virt.hardware [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 698.334645] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70ff8e4-eb68-417a-a059-a9c84a73808f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.343896] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ba23b7-0a3e-4b0e-84a7-94c89518c61a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.421127] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "225c02d6-43df-4260-a428-14a6b1ed32cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.421765] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.481567] env[70013]: DEBUG nova.network.neutron [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Successfully created port: 63c17eba-d2d2-44c2-bd66-bcbc6d27e157 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.510721] env[70013]: DEBUG nova.network.neutron [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Updating instance_info_cache with network_info: [{"id": "ecf33e4a-0e1b-41fe-b723-bd679a50d3a0", "address": "fa:16:3e:35:df:94", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf33e4a-0e", "ovs_interfaceid": "ecf33e4a-0e1b-41fe-b723-bd679a50d3a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.536065] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.284s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.536122] env[70013]: DEBUG nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 698.539032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.184s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.540541] env[70013]: INFO nova.compute.claims [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.701419] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027855} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.701626] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 698.702392] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-887ca4d5-a6fe-44b8-a87d-afdc96580e23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.713862] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 698.713862] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e37471-569e-283d-4443-e6b80c733ffb" [ 698.713862] env[70013]: _type = "Task" [ 698.713862] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.726138] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e37471-569e-283d-4443-e6b80c733ffb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.925104] env[70013]: DEBUG nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 699.017592] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Releasing lock "refresh_cache-2b91ec52-2d32-4f86-acf2-f74f661f65b5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.018126] env[70013]: DEBUG nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Instance network_info: |[{"id": "ecf33e4a-0e1b-41fe-b723-bd679a50d3a0", "address": "fa:16:3e:35:df:94", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf33e4a-0e", "ovs_interfaceid": "ecf33e4a-0e1b-41fe-b723-bd679a50d3a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 699.018600] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:df:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ecf33e4a-0e1b-41fe-b723-bd679a50d3a0', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.028045] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Creating folder: Project (7691ed6dbc544b6e892c72a4235deac8). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.029407] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b9a3dd5-1199-4491-ae37-3026354ab3aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.041786] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Created folder: Project (7691ed6dbc544b6e892c72a4235deac8) in parent group-v836999. [ 699.042086] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Creating folder: Instances. Parent ref: group-v837003. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.042902] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f30d36ae-b7f4-4e3c-b616-5dc74977ed0d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.047441] env[70013]: DEBUG nova.compute.utils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 699.061273] env[70013]: DEBUG nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 699.061273] env[70013]: DEBUG nova.network.neutron [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.077368] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Created folder: Instances in parent group-v837003. [ 699.077746] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 699.078034] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 699.078336] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-793ff8b2-763e-4617-b799-c40624c7cbdb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.101723] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.101723] env[70013]: value = "task-4230467" [ 699.101723] env[70013]: _type = "Task" [ 699.101723] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.111702] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230467, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.141924] env[70013]: DEBUG nova.policy [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b58fea544a4b4504855231d8da4f11b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e7f2ac8bf549e1ad6b804da80364f9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 699.231526] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e37471-569e-283d-4443-e6b80c733ffb, 'name': SearchDatastore_Task, 'duration_secs': 0.008405} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.231833] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.232124] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.232410] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b102a0d-520d-4211-9102-26b08365c057 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.243040] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 699.243040] env[70013]: value = "task-4230468" [ 699.243040] env[70013]: _type = "Task" [ 699.243040] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.253589] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230468, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.465619] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.561776] env[70013]: DEBUG nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 699.584234] env[70013]: DEBUG nova.network.neutron [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Successfully created port: 7edceb22-219c-4108-abc8-dabed5a1e064 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.615429] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230467, 'name': CreateVM_Task, 'duration_secs': 0.467355} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.618375] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 699.638291] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.638756] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.638901] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 699.640397] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a972d81e-2f69-425e-9b85-50f8d625394e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.648499] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 699.648499] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521563ee-e6a6-2aaa-79b8-d40233a6399d" [ 699.648499] env[70013]: _type = "Task" [ 699.648499] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.658434] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521563ee-e6a6-2aaa-79b8-d40233a6399d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.755040] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230468, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506897} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.757908] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 699.758179] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 699.759953] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8df11195-8375-4b77-b428-7c471bedf5da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.768833] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 699.768833] env[70013]: value = "task-4230469" [ 699.768833] env[70013]: _type = "Task" [ 699.768833] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.784697] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.874613] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbaa39d-3b18-4c4b-ae75-3315e286b939 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.883935] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0ee1ea-0ab3-4480-a016-716edd0091d4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.921452] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b4037f-a7f9-49a3-853f-c235a01e5e13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.930664] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d124cdcd-d29e-426b-a6e6-ae50ada0b238 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.950605] env[70013]: DEBUG nova.compute.provider_tree [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.162031] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521563ee-e6a6-2aaa-79b8-d40233a6399d, 'name': SearchDatastore_Task, 'duration_secs': 0.061188} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.162365] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.162581] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.162764] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.162916] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.163097] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.163376] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4301ca4-ba94-488d-9560-9354ca64d062 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.173951] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.173951] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 700.175728] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3ddfebf-c1d1-4afd-a236-6ba60361f81f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.181531] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 700.181531] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52eebd10-63bc-5ba6-e593-46230026150c" [ 700.181531] env[70013]: _type = "Task" [ 700.181531] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.203058] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52eebd10-63bc-5ba6-e593-46230026150c, 'name': SearchDatastore_Task, 'duration_secs': 0.011454} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.203414] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5ca86dc-fe19-4eae-8551-f3676fed1eda {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.213812] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 700.213812] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5208090f-32a8-9eed-a9d2-ab1c23408786" [ 700.213812] env[70013]: _type = "Task" [ 700.213812] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.226023] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5208090f-32a8-9eed-a9d2-ab1c23408786, 'name': SearchDatastore_Task, 'duration_secs': 0.009281} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.226304] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.226566] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2b91ec52-2d32-4f86-acf2-f74f661f65b5/2b91ec52-2d32-4f86-acf2-f74f661f65b5.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 700.226840] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cc79d83-61f3-4775-b9e9-3b1ecb8d4e07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.237603] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 700.237603] env[70013]: value = "task-4230470" [ 700.237603] env[70013]: _type = "Task" [ 700.237603] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.247052] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.283261] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230469, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072812} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.283261] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 700.284665] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058e9e80-400e-4cbf-986c-336287c3693f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.313262] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 700.314052] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a765c033-8448-4ee7-9778-a6aa9a2b3ae4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.346185] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 700.346185] env[70013]: value = "task-4230471" [ 700.346185] env[70013]: _type = "Task" [ 700.346185] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.354123] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230471, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.457834] env[70013]: DEBUG nova.scheduler.client.report [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 700.579170] env[70013]: DEBUG nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 700.618342] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 700.618607] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.618607] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 700.618800] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.619168] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 700.619168] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 700.619979] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 700.619979] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 700.619979] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 700.619979] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 700.619979] env[70013]: DEBUG nova.virt.hardware [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 700.621080] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e959f8d1-5c5b-435a-b684-f3400297443a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.632210] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c3f984-adc6-43cd-b798-e0fa020cd755 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.750126] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230470, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502431} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.750610] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2b91ec52-2d32-4f86-acf2-f74f661f65b5/2b91ec52-2d32-4f86-acf2-f74f661f65b5.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 700.751050] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.751512] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d584fd6-63fc-4bea-81e1-b531596d78df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.762528] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 700.762528] env[70013]: value = "task-4230472" [ 700.762528] env[70013]: _type = "Task" [ 700.762528] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.773377] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230472, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.858873] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230471, 'name': ReconfigVM_Task, 'duration_secs': 0.509677} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.859040] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 700.859782] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9250c211-9471-4cbd-aeb1-1e84344d3065 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.868193] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 700.868193] env[70013]: value = "task-4230473" [ 700.868193] env[70013]: _type = "Task" [ 700.868193] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.880378] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230473, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.964195] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 700.965234] env[70013]: DEBUG nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 700.971470] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.835s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.973349] env[70013]: INFO nova.compute.claims [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.004327] env[70013]: DEBUG nova.network.neutron [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Successfully updated port: 63c17eba-d2d2-44c2-bd66-bcbc6d27e157 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 701.279042] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230472, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077984} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.279042] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.280333] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2ed257-8e48-4f21-b6c9-1afdefdcecf3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.312972] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 2b91ec52-2d32-4f86-acf2-f74f661f65b5/2b91ec52-2d32-4f86-acf2-f74f661f65b5.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.313239] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fcde0b6f-d07e-4244-85de-21d7bea10b83 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.333863] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 701.333863] env[70013]: value = "task-4230474" [ 701.333863] env[70013]: _type = "Task" [ 701.333863] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.346620] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230474, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.382544] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230473, 'name': Rename_Task, 'duration_secs': 0.176986} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.382856] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 701.383458] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d6def5a-7f42-460e-b101-f181eb8f74a0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.394304] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 701.394304] env[70013]: value = "task-4230475" [ 701.394304] env[70013]: _type = "Task" [ 701.394304] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.406472] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.473104] env[70013]: DEBUG nova.compute.utils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 701.477191] env[70013]: DEBUG nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 701.477932] env[70013]: DEBUG nova.network.neutron [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 701.507261] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "refresh_cache-e8998cab-533b-46b1-a1ea-5eb5db7d6d99" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.507444] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired lock "refresh_cache-e8998cab-533b-46b1-a1ea-5eb5db7d6d99" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.507627] env[70013]: DEBUG nova.network.neutron [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.673326] env[70013]: DEBUG nova.policy [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a47dea2686244b5e86cdd6a7cbe8a4c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8214085b60d0482596cdb3e8df48c837', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 701.850487] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230474, 'name': ReconfigVM_Task, 'duration_secs': 0.399618} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.850487] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 2b91ec52-2d32-4f86-acf2-f74f661f65b5/2b91ec52-2d32-4f86-acf2-f74f661f65b5.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 701.850487] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fb4ca9c-0a1a-42a2-b00c-13b22ce6d3b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.853029] env[70013]: DEBUG nova.network.neutron [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Successfully updated port: 7edceb22-219c-4108-abc8-dabed5a1e064 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 701.861762] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 701.861762] env[70013]: value = "task-4230476" [ 701.861762] env[70013]: _type = "Task" [ 701.861762] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.874342] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230476, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.906335] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230475, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.985393] env[70013]: DEBUG nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 702.093691] env[70013]: DEBUG nova.network.neutron [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.151538] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "63c3b57c-022a-4eee-b215-6dd89142e659" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.151767] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.313656] env[70013]: DEBUG nova.compute.manager [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Received event network-changed-ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 702.313906] env[70013]: DEBUG nova.compute.manager [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Refreshing instance network info cache due to event network-changed-ecf33e4a-0e1b-41fe-b723-bd679a50d3a0. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 702.314082] env[70013]: DEBUG oslo_concurrency.lockutils [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] Acquiring lock "refresh_cache-2b91ec52-2d32-4f86-acf2-f74f661f65b5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.314228] env[70013]: DEBUG oslo_concurrency.lockutils [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] Acquired lock "refresh_cache-2b91ec52-2d32-4f86-acf2-f74f661f65b5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.314386] env[70013]: DEBUG nova.network.neutron [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Refreshing network info cache for port ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 702.356809] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "refresh_cache-9c5f2422-820e-454b-98be-695e3ac6d1e3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.356979] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquired lock "refresh_cache-9c5f2422-820e-454b-98be-695e3ac6d1e3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.357163] env[70013]: DEBUG nova.network.neutron [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.373555] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edcb57b-df04-4879-9745-547aeb7cd4a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.382446] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230476, 'name': Rename_Task, 'duration_secs': 0.412707} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.383207] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.383800] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5f05839-c52c-4538-9f37-e6fad2dadd96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.391947] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 702.391947] env[70013]: value = "task-4230477" [ 702.391947] env[70013]: _type = "Task" [ 702.391947] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.399420] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a682bf7-7619-4b15-bcee-f68599f2e3df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.414623] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230477, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.449619] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb04e23-f473-4632-afc2-0c1159be1b47 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.456423] env[70013]: DEBUG oslo_vmware.api [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230475, 'name': PowerOnVM_Task, 'duration_secs': 0.520498} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.456800] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 702.457267] env[70013]: INFO nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Took 8.83 seconds to spawn the instance on the hypervisor. [ 702.457534] env[70013]: DEBUG nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 702.461491] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae9832a-9dda-401a-ac70-8344a589326f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.464675] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00bfb472-1112-4c65-b6de-6cc7a796c10e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.485863] env[70013]: DEBUG nova.compute.provider_tree [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.563650] env[70013]: DEBUG nova.network.neutron [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Updating instance_info_cache with network_info: [{"id": "63c17eba-d2d2-44c2-bd66-bcbc6d27e157", "address": "fa:16:3e:c2:ed:df", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63c17eba-d2", "ovs_interfaceid": "63c17eba-d2d2-44c2-bd66-bcbc6d27e157", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.908964] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230477, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.952186] env[70013]: DEBUG nova.network.neutron [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.983786] env[70013]: INFO nova.compute.manager [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Took 14.64 seconds to build instance. [ 702.996025] env[70013]: DEBUG nova.scheduler.client.report [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 702.999937] env[70013]: DEBUG nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 703.054521] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 703.054521] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.054521] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 703.055876] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.055876] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 703.055876] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 703.055876] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 703.055876] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 703.056024] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 703.056024] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 703.056024] env[70013]: DEBUG nova.virt.hardware [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 703.056024] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6386fa53-9778-4ffa-b862-2d197526d577 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.065418] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08612cb-92e4-47e8-96e1-647a645cf706 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.071226] env[70013]: DEBUG nova.network.neutron [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Successfully created port: 21a0b03c-bc07-44c7-b9e4-503338c106cd {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 703.076525] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Releasing lock "refresh_cache-e8998cab-533b-46b1-a1ea-5eb5db7d6d99" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.076979] env[70013]: DEBUG nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Instance network_info: |[{"id": "63c17eba-d2d2-44c2-bd66-bcbc6d27e157", "address": "fa:16:3e:c2:ed:df", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63c17eba-d2", "ovs_interfaceid": "63c17eba-d2d2-44c2-bd66-bcbc6d27e157", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 703.077922] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:ed:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63c17eba-d2d2-44c2-bd66-bcbc6d27e157', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 703.090045] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Creating folder: Project (dbdf4732a346424b8ff1c39c7e900f08). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.091524] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-573912fe-45a1-466b-ba5c-d200243ab017 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.118396] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Created folder: Project (dbdf4732a346424b8ff1c39c7e900f08) in parent group-v836999. [ 703.118396] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Creating folder: Instances. Parent ref: group-v837006. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.118396] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd064059-6ef9-47bb-b8c3-7cb91a1c5c0d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.128871] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Created folder: Instances in parent group-v837006. [ 703.129149] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 703.129392] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 703.129626] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4555991c-d032-4a8c-91fa-293bcc83fc9a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.152987] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 703.152987] env[70013]: value = "task-4230480" [ 703.152987] env[70013]: _type = "Task" [ 703.152987] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.161755] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230480, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.412976] env[70013]: DEBUG oslo_vmware.api [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230477, 'name': PowerOnVM_Task, 'duration_secs': 0.987781} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.414616] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 703.414824] env[70013]: INFO nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Took 11.98 seconds to spawn the instance on the hypervisor. [ 703.415066] env[70013]: DEBUG nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 703.416350] env[70013]: DEBUG nova.compute.manager [req-27a02d35-17da-4131-882a-e68695f7f296 req-b08a5264-5679-46e9-ab4b-d331be74fa82 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Received event network-vif-plugged-7edceb22-219c-4108-abc8-dabed5a1e064 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 703.416547] env[70013]: DEBUG oslo_concurrency.lockutils [req-27a02d35-17da-4131-882a-e68695f7f296 req-b08a5264-5679-46e9-ab4b-d331be74fa82 service nova] Acquiring lock "9c5f2422-820e-454b-98be-695e3ac6d1e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.416740] env[70013]: DEBUG oslo_concurrency.lockutils [req-27a02d35-17da-4131-882a-e68695f7f296 req-b08a5264-5679-46e9-ab4b-d331be74fa82 service nova] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.416896] env[70013]: DEBUG oslo_concurrency.lockutils [req-27a02d35-17da-4131-882a-e68695f7f296 req-b08a5264-5679-46e9-ab4b-d331be74fa82 service nova] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.417067] env[70013]: DEBUG nova.compute.manager [req-27a02d35-17da-4131-882a-e68695f7f296 req-b08a5264-5679-46e9-ab4b-d331be74fa82 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] No waiting events found dispatching network-vif-plugged-7edceb22-219c-4108-abc8-dabed5a1e064 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 703.417233] env[70013]: WARNING nova.compute.manager [req-27a02d35-17da-4131-882a-e68695f7f296 req-b08a5264-5679-46e9-ab4b-d331be74fa82 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Received unexpected event network-vif-plugged-7edceb22-219c-4108-abc8-dabed5a1e064 for instance with vm_state building and task_state spawning. [ 703.419392] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba15ab3b-9270-4fb0-a2f2-f6509d0e7d84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.461775] env[70013]: DEBUG nova.network.neutron [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Updating instance_info_cache with network_info: [{"id": "7edceb22-219c-4108-abc8-dabed5a1e064", "address": "fa:16:3e:85:bc:70", "network": {"id": "99863cb4-1242-4006-8228-f6d29ea7203b", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1119228590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e7f2ac8bf549e1ad6b804da80364f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7edceb22-21", "ovs_interfaceid": "7edceb22-219c-4108-abc8-dabed5a1e064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.490671] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2317eec4-856e-4907-8658-90b6c8c1937e tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.166s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.492437] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.771s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.492629] env[70013]: INFO nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] During sync_power_state the instance has a pending task (spawning). Skip. [ 703.492892] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.508286] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.509049] env[70013]: DEBUG nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 703.515190] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.241s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.517684] env[70013]: INFO nova.compute.claims [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.663319] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230480, 'name': CreateVM_Task, 'duration_secs': 0.415561} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.663550] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 703.664200] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.664985] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.664985] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 703.666583] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4b86693-0f62-4264-a40b-95907e07ea78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.672121] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 703.672121] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5228e796-d071-d66b-0519-3d5671933cb1" [ 703.672121] env[70013]: _type = "Task" [ 703.672121] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.683223] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5228e796-d071-d66b-0519-3d5671933cb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.867212] env[70013]: DEBUG nova.network.neutron [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Updated VIF entry in instance network info cache for port ecf33e4a-0e1b-41fe-b723-bd679a50d3a0. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 703.867795] env[70013]: DEBUG nova.network.neutron [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Updating instance_info_cache with network_info: [{"id": "ecf33e4a-0e1b-41fe-b723-bd679a50d3a0", "address": "fa:16:3e:35:df:94", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf33e4a-0e", "ovs_interfaceid": "ecf33e4a-0e1b-41fe-b723-bd679a50d3a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.953632] env[70013]: INFO nova.compute.manager [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Took 16.80 seconds to build instance. [ 703.968799] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Releasing lock "refresh_cache-9c5f2422-820e-454b-98be-695e3ac6d1e3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.969160] env[70013]: DEBUG nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Instance network_info: |[{"id": "7edceb22-219c-4108-abc8-dabed5a1e064", "address": "fa:16:3e:85:bc:70", "network": {"id": "99863cb4-1242-4006-8228-f6d29ea7203b", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1119228590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e7f2ac8bf549e1ad6b804da80364f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7edceb22-21", "ovs_interfaceid": "7edceb22-219c-4108-abc8-dabed5a1e064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 703.969980] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:bc:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7edceb22-219c-4108-abc8-dabed5a1e064', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 703.978860] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Creating folder: Project (03e7f2ac8bf549e1ad6b804da80364f9). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.980074] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1e0954c-dd56-4d65-8b65-16e128656cd9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.993971] env[70013]: DEBUG nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 703.997701] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Created folder: Project (03e7f2ac8bf549e1ad6b804da80364f9) in parent group-v836999. [ 703.998106] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Creating folder: Instances. Parent ref: group-v837009. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.998463] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5fd26eed-7008-457e-8d96-b9f175e98fc2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.009444] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Created folder: Instances in parent group-v837009. [ 704.009705] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 704.010339] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 704.010607] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb098f17-0897-4195-9b00-2fb269d1c567 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.036470] env[70013]: DEBUG nova.compute.utils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 704.045324] env[70013]: DEBUG nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 704.045661] env[70013]: DEBUG nova.network.neutron [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 704.055310] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.055310] env[70013]: value = "task-4230483" [ 704.055310] env[70013]: _type = "Task" [ 704.055310] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.067904] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230483, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.191554] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5228e796-d071-d66b-0519-3d5671933cb1, 'name': SearchDatastore_Task, 'duration_secs': 0.009819} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.191554] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.191554] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.191554] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.191848] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.191991] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.192287] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-350b5631-26da-4181-aa5a-a4905895651f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.201920] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.202065] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 704.203177] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9063abdd-c1f1-48e9-b939-6b4c673e7853 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.212955] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 704.212955] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1a1a4-fef4-e702-9bf7-7a3cc6b1a471" [ 704.212955] env[70013]: _type = "Task" [ 704.212955] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.228793] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1a1a4-fef4-e702-9bf7-7a3cc6b1a471, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.321181] env[70013]: DEBUG nova.policy [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9656269471843f3a0cad8e5a2cfe296', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f00f00d83424b5b8abfd6b22d9e119b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 704.371109] env[70013]: DEBUG oslo_concurrency.lockutils [req-996c22c7-bf39-4e8f-b308-21013e0d014c req-94d87d98-fb6c-4afc-98c6-d1ad63208bf7 service nova] Releasing lock "refresh_cache-2b91ec52-2d32-4f86-acf2-f74f661f65b5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.458698] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae7e64fc-630f-4bcc-8823-c631207526f0 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.318s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.460720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.738s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.460720] env[70013]: INFO nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] During sync_power_state the instance has a pending task (spawning). Skip. [ 704.460720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.540876] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.546167] env[70013]: DEBUG nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 704.571171] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230483, 'name': CreateVM_Task, 'duration_secs': 0.455351} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.573073] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 704.574465] env[70013]: DEBUG oslo_vmware.service [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3347b478-2389-4b80-a703-5f1918d2b2b2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.590050] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.590329] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.591491] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 704.591491] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e79f3b9b-4a2d-46bd-8b98-df15b0a08f50 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.605107] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 704.605107] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a8e9da-d091-20da-7fd2-539cebbf5169" [ 704.605107] env[70013]: _type = "Task" [ 704.605107] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.618682] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.618894] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.619158] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.619756] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.619756] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.620195] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d4168b7-576a-4e2c-91d1-f5935fe4eadc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.632187] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.632187] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 704.632953] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ff3cfa-6789-46c0-b6f0-9084e62efeb1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.642666] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16ab1e9b-544b-4f41-97ed-d93a780f943f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.650687] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 704.650687] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5246d02a-3eeb-f975-fe16-407ecfa76906" [ 704.650687] env[70013]: _type = "Task" [ 704.650687] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.668606] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Preparing fetch location {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 704.669299] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Creating directory with path [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.669647] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a5fcb49-a9fa-48a6-9b69-61541cb86f5f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.695243] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Created directory with path [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.695622] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Fetch image to [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 704.696910] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Downloading image file data 0ddeef65-a14c-4065-97bb-58b607968d19 to [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk on the data store datastore1 {{(pid=70013) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 704.697935] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6144d0c5-94f9-4447-950c-eea29b268f75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.706915] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d7bfd1-f9d0-4bfa-9acd-eb85ef545c8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.727785] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "64308d7c-b63b-4fa6-b235-2f0586a888cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.728081] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.741837] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eae528e-801c-4611-aade-ab3fec3c4f7d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.747208] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1a1a4-fef4-e702-9bf7-7a3cc6b1a471, 'name': SearchDatastore_Task, 'duration_secs': 0.021894} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.753516] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60a96dfd-4336-4b54-9751-77713ce7d4f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.793946] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0770dc-56cc-42ae-8a2d-181bd97437c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.799583] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 704.799583] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52918002-6f5a-5245-96ae-111659e496ba" [ 704.799583] env[70013]: _type = "Task" [ 704.799583] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.806454] env[70013]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-5f75d2e1-fa0a-4253-a8af-1ee0d890e10e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.813029] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52918002-6f5a-5245-96ae-111659e496ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.846583] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Downloading image file data 0ddeef65-a14c-4065-97bb-58b607968d19 to the data store datastore1 {{(pid=70013) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 704.909368] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e551d84-f7cf-4e5e-820f-4a3eb7619687 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.923562] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535c91ae-dfd8-4035-a477-7512ff7a0270 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.974076] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=70013) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 704.975852] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea6c5e5-2718-4051-948d-3aa2c602b00f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.038376] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1121750b-169f-4c8f-9588-eb5af9fcf953 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.069036] env[70013]: DEBUG nova.compute.provider_tree [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.134516] env[70013]: DEBUG nova.network.neutron [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Successfully created port: b402fce4-7392-4139-8de5-7620820e0814 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.210011] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.210344] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.241291] env[70013]: DEBUG nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 705.315154] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52918002-6f5a-5245-96ae-111659e496ba, 'name': SearchDatastore_Task, 'duration_secs': 0.018464} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.319201] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.319412] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e8998cab-533b-46b1-a1ea-5eb5db7d6d99/e8998cab-533b-46b1-a1ea-5eb5db7d6d99.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.319592] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-364a616f-88ca-4c23-a7cc-c1555ed5cecf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.328582] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 705.328582] env[70013]: value = "task-4230484" [ 705.328582] env[70013]: _type = "Task" [ 705.328582] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.340762] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230484, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.574865] env[70013]: DEBUG nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 705.577844] env[70013]: DEBUG nova.scheduler.client.report [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 705.616250] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 705.616250] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.616250] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 705.616250] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.616524] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 705.616524] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 705.616524] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 705.616524] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 705.616524] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 705.616715] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 705.616715] env[70013]: DEBUG nova.virt.hardware [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 705.621503] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3840ee6b-f733-4ef0-96bc-7bff293d85c6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.633036] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f060937-6d4c-4fee-8caf-f1b87bce3994 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.781412] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.842188] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230484, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.857969] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Completed reading data from the image iterator. {{(pid=70013) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 705.857969] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 705.912198] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Downloaded image file data 0ddeef65-a14c-4065-97bb-58b607968d19 to vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk on the data store datastore1 {{(pid=70013) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 705.913961] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Caching image {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 705.914260] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Copying Virtual Disk [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk to [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.914549] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f668a09-28c3-4a36-9bbc-8e53ecde7fb4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.923962] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 705.923962] env[70013]: value = "task-4230485" [ 705.923962] env[70013]: _type = "Task" [ 705.923962] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.935178] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230485, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.088165] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.088165] env[70013]: DEBUG nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 706.090636] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.445s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.092493] env[70013]: INFO nova.compute.claims [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.154983] env[70013]: DEBUG nova.compute.manager [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Received event network-vif-plugged-63c17eba-d2d2-44c2-bd66-bcbc6d27e157 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 706.154983] env[70013]: DEBUG oslo_concurrency.lockutils [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] Acquiring lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.154983] env[70013]: DEBUG oslo_concurrency.lockutils [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.154983] env[70013]: DEBUG oslo_concurrency.lockutils [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.154983] env[70013]: DEBUG nova.compute.manager [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] No waiting events found dispatching network-vif-plugged-63c17eba-d2d2-44c2-bd66-bcbc6d27e157 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 706.155188] env[70013]: WARNING nova.compute.manager [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Received unexpected event network-vif-plugged-63c17eba-d2d2-44c2-bd66-bcbc6d27e157 for instance with vm_state building and task_state spawning. [ 706.155188] env[70013]: DEBUG nova.compute.manager [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Received event network-changed-63c17eba-d2d2-44c2-bd66-bcbc6d27e157 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 706.155188] env[70013]: DEBUG nova.compute.manager [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Refreshing instance network info cache due to event network-changed-63c17eba-d2d2-44c2-bd66-bcbc6d27e157. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 706.155371] env[70013]: DEBUG oslo_concurrency.lockutils [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] Acquiring lock "refresh_cache-e8998cab-533b-46b1-a1ea-5eb5db7d6d99" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.155512] env[70013]: DEBUG oslo_concurrency.lockutils [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] Acquired lock "refresh_cache-e8998cab-533b-46b1-a1ea-5eb5db7d6d99" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 706.155693] env[70013]: DEBUG nova.network.neutron [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Refreshing network info cache for port 63c17eba-d2d2-44c2-bd66-bcbc6d27e157 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 706.187531] env[70013]: INFO nova.compute.manager [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Rebuilding instance [ 706.280038] env[70013]: DEBUG nova.compute.manager [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 706.280038] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7852ac-503a-41cd-93ff-5aaf10cb9f4d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.341905] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230484, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550138} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.343051] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e8998cab-533b-46b1-a1ea-5eb5db7d6d99/e8998cab-533b-46b1-a1ea-5eb5db7d6d99.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.343051] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.343051] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-32f9968c-9022-466c-afe3-55db98cabaea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.351375] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 706.351375] env[70013]: value = "task-4230486" [ 706.351375] env[70013]: _type = "Task" [ 706.351375] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.363388] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230486, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.395320] env[70013]: DEBUG nova.network.neutron [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Successfully updated port: 21a0b03c-bc07-44c7-b9e4-503338c106cd {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 706.447343] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230485, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.516437] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "fc0b7f87-8a36-440e-b158-30b9d9920572" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.518038] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.597996] env[70013]: DEBUG nova.compute.utils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 706.602477] env[70013]: DEBUG nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 706.602613] env[70013]: DEBUG nova.network.neutron [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 706.778180] env[70013]: DEBUG nova.policy [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '523a56f27ffe4c50bd1d63fc6458b291', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b4f41a91a5d4b679adafa6624a1db74', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 706.872841] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230486, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076764} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.872841] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 706.872841] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91baaca2-8866-4b83-95f4-a970b1719309 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.908219] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] e8998cab-533b-46b1-a1ea-5eb5db7d6d99/e8998cab-533b-46b1-a1ea-5eb5db7d6d99.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 706.909142] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "refresh_cache-417f8c74-81cc-47d6-80ca-b61d03ebec50" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.909142] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquired lock "refresh_cache-417f8c74-81cc-47d6-80ca-b61d03ebec50" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 706.909285] env[70013]: DEBUG nova.network.neutron [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.911355] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ade32629-aa85-477e-b1e0-8016736a3f5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.936021] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 706.936021] env[70013]: value = "task-4230487" [ 706.936021] env[70013]: _type = "Task" [ 706.936021] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.951570] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230485, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.728589} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.952830] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Copied Virtual Disk [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk to [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.953225] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Deleting the datastore file [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19/tmp-sparse.vmdk {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.953326] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de414c5e-1d8c-4111-a7ae-11c96d4909c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.958870] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230487, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.965972] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 706.965972] env[70013]: value = "task-4230488" [ 706.965972] env[70013]: _type = "Task" [ 706.965972] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.974156] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230488, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.106009] env[70013]: DEBUG nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 707.184432] env[70013]: DEBUG nova.compute.manager [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Received event network-changed-7edceb22-219c-4108-abc8-dabed5a1e064 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 707.184669] env[70013]: DEBUG nova.compute.manager [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Refreshing instance network info cache due to event network-changed-7edceb22-219c-4108-abc8-dabed5a1e064. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 707.187181] env[70013]: DEBUG oslo_concurrency.lockutils [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] Acquiring lock "refresh_cache-9c5f2422-820e-454b-98be-695e3ac6d1e3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.187181] env[70013]: DEBUG oslo_concurrency.lockutils [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] Acquired lock "refresh_cache-9c5f2422-820e-454b-98be-695e3ac6d1e3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.187181] env[70013]: DEBUG nova.network.neutron [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Refreshing network info cache for port 7edceb22-219c-4108-abc8-dabed5a1e064 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.308237] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 707.308237] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee08d603-29c7-4163-8c83-12f91fa46cd4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.311276] env[70013]: DEBUG nova.network.neutron [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Successfully updated port: b402fce4-7392-4139-8de5-7620820e0814 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 707.315336] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 707.315336] env[70013]: value = "task-4230489" [ 707.315336] env[70013]: _type = "Task" [ 707.315336] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.334919] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.451448] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230487, 'name': ReconfigVM_Task, 'duration_secs': 0.345702} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.455066] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Reconfigured VM instance instance-00000003 to attach disk [datastore2] e8998cab-533b-46b1-a1ea-5eb5db7d6d99/e8998cab-533b-46b1-a1ea-5eb5db7d6d99.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.456693] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71a71896-84e0-4a0a-9e87-e2b7d7a72eec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.466859] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 707.466859] env[70013]: value = "task-4230490" [ 707.466859] env[70013]: _type = "Task" [ 707.466859] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.495286] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230490, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.495646] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230488, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028392} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.498276] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.498276] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Moving file from [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285/0ddeef65-a14c-4065-97bb-58b607968d19 to [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19. {{(pid=70013) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 707.499065] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe46913-04b3-4585-92c7-0b8982ef389d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.505095] env[70013]: DEBUG nova.network.neutron [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.507499] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-6854bcff-f664-4d69-9bc9-87ae9c7c697e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.518889] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd729fb-ca97-4a1e-8a50-be020f995695 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.522951] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 707.522951] env[70013]: value = "task-4230491" [ 707.522951] env[70013]: _type = "Task" [ 707.522951] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.565775] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43f62da-62cd-46b5-a406-820d9d028b39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.574529] env[70013]: DEBUG nova.network.neutron [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Updated VIF entry in instance network info cache for port 63c17eba-d2d2-44c2-bd66-bcbc6d27e157. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 707.574529] env[70013]: DEBUG nova.network.neutron [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Updating instance_info_cache with network_info: [{"id": "63c17eba-d2d2-44c2-bd66-bcbc6d27e157", "address": "fa:16:3e:c2:ed:df", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63c17eba-d2", "ovs_interfaceid": "63c17eba-d2d2-44c2-bd66-bcbc6d27e157", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.579495] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230491, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029822} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.580505] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] File moved {{(pid=70013) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 707.581416] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Cleaning up location [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 707.581416] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Deleting the datastore file [datastore1] vmware_temp/510aad71-77a4-4bc2-99ef-250ef3a83285 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 707.581416] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fea1e597-80b5-4681-8ead-4a9611b6a860 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.587614] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4e14b7-796e-43b7-a097-1951a7ca88e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.595112] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 707.595112] env[70013]: value = "task-4230492" [ 707.595112] env[70013]: _type = "Task" [ 707.595112] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.617833] env[70013]: DEBUG nova.compute.provider_tree [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.626537] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230492, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.814731] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.814731] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.814731] env[70013]: DEBUG nova.network.neutron [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.832642] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230489, 'name': PowerOffVM_Task, 'duration_secs': 0.312439} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.833578] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 707.833825] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.834667] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5e978a-3d17-4c66-8ff1-6468f46db395 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.844490] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 707.845211] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60536eea-d5c9-44f5-9be5-659b1bc24095 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.874063] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 707.874063] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 707.874218] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleting the datastore file [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 707.874486] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cc2db02-90eb-44ab-8cb4-6e9c64a54a89 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.884134] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 707.884134] env[70013]: value = "task-4230494" [ 707.884134] env[70013]: _type = "Task" [ 707.884134] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.893266] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230494, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.984616] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230490, 'name': Rename_Task, 'duration_secs': 0.189213} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.984938] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 707.985216] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a5a07c7-50ef-4873-9b66-c27234450152 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.993477] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 707.993477] env[70013]: value = "task-4230495" [ 707.993477] env[70013]: _type = "Task" [ 707.993477] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.003923] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230495, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.005272] env[70013]: DEBUG nova.network.neutron [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Updating instance_info_cache with network_info: [{"id": "21a0b03c-bc07-44c7-b9e4-503338c106cd", "address": "fa:16:3e:8e:9d:ae", "network": {"id": "8aa548f3-6ae4-440d-8d55-f22e41739121", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1100728332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8214085b60d0482596cdb3e8df48c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21a0b03c-bc", "ovs_interfaceid": "21a0b03c-bc07-44c7-b9e4-503338c106cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.081430] env[70013]: DEBUG oslo_concurrency.lockutils [req-cb2d4892-6ce9-42c5-9e98-cb6a6af68349 req-26f2e156-5370-42c7-9753-b48dfef1d5d2 service nova] Releasing lock "refresh_cache-e8998cab-533b-46b1-a1ea-5eb5db7d6d99" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.108936] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230492, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042029} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.109232] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.110157] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7003ec0b-5f58-4e5f-9c03-8158fc782447 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.117665] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 708.117665] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5227f8c9-278b-7cf9-bbac-b1f11abe0d07" [ 708.117665] env[70013]: _type = "Task" [ 708.117665] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.124373] env[70013]: DEBUG nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 708.127137] env[70013]: DEBUG nova.scheduler.client.report [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 708.139258] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5227f8c9-278b-7cf9-bbac-b1f11abe0d07, 'name': SearchDatastore_Task, 'duration_secs': 0.01052} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.139635] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.140433] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9c5f2422-820e-454b-98be-695e3ac6d1e3/9c5f2422-820e-454b-98be-695e3ac6d1e3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 708.141028] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a4e180a-76a8-4e67-b889-12b5bf80b35b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.152021] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 708.152021] env[70013]: value = "task-4230496" [ 708.152021] env[70013]: _type = "Task" [ 708.152021] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.170094] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230496, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.172659] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 708.172886] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.173054] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 708.173244] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.173389] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 708.173536] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 708.173777] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 708.174012] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 708.174233] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 708.174406] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 708.174596] env[70013]: DEBUG nova.virt.hardware [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 708.175785] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab216e8-4642-4607-9385-177e220c4b20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.185336] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251c636f-7102-4c30-84a5-db17fa94280b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.240237] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "fc473d40-b57b-437e-9511-58a0ba700a69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.240488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.381348] env[70013]: DEBUG nova.network.neutron [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.399515] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230494, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101072} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.401124] env[70013]: DEBUG nova.network.neutron [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Successfully created port: c9bdfb0a-f6f0-4467-b641-a8c04598910f {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.402843] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.402843] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 708.402992] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.511114] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Releasing lock "refresh_cache-417f8c74-81cc-47d6-80ca-b61d03ebec50" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.511353] env[70013]: DEBUG nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Instance network_info: |[{"id": "21a0b03c-bc07-44c7-b9e4-503338c106cd", "address": "fa:16:3e:8e:9d:ae", "network": {"id": "8aa548f3-6ae4-440d-8d55-f22e41739121", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1100728332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8214085b60d0482596cdb3e8df48c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21a0b03c-bc", "ovs_interfaceid": "21a0b03c-bc07-44c7-b9e4-503338c106cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 708.512245] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230495, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.515761] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:9d:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '77aa121f-8fb6-42f3-aaea-43addfe449b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21a0b03c-bc07-44c7-b9e4-503338c106cd', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.524957] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Creating folder: Project (8214085b60d0482596cdb3e8df48c837). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.525175] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2e781192-20f0-4eeb-b20e-7a6a7f4f2df3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.539841] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Created folder: Project (8214085b60d0482596cdb3e8df48c837) in parent group-v836999. [ 708.540088] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Creating folder: Instances. Parent ref: group-v837012. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.540413] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82968292-595d-47bd-90a8-3a14098ef618 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.553594] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Created folder: Instances in parent group-v837012. [ 708.553909] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 708.556188] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 708.556188] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6803c6d1-03b1-4215-80cc-1b7cfc79b4a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.589326] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.589326] env[70013]: value = "task-4230499" [ 708.589326] env[70013]: _type = "Task" [ 708.589326] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.610165] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230499, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.633468] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 708.634422] env[70013]: DEBUG nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 708.637713] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.955s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.640951] env[70013]: INFO nova.compute.claims [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.667534] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230496, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.753841] env[70013]: DEBUG nova.network.neutron [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance_info_cache with network_info: [{"id": "b402fce4-7392-4139-8de5-7620820e0814", "address": "fa:16:3e:44:ef:b6", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb402fce4-73", "ovs_interfaceid": "b402fce4-7392-4139-8de5-7620820e0814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.895068] env[70013]: DEBUG nova.network.neutron [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Updated VIF entry in instance network info cache for port 7edceb22-219c-4108-abc8-dabed5a1e064. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 708.895068] env[70013]: DEBUG nova.network.neutron [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Updating instance_info_cache with network_info: [{"id": "7edceb22-219c-4108-abc8-dabed5a1e064", "address": "fa:16:3e:85:bc:70", "network": {"id": "99863cb4-1242-4006-8228-f6d29ea7203b", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1119228590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e7f2ac8bf549e1ad6b804da80364f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7edceb22-21", "ovs_interfaceid": "7edceb22-219c-4108-abc8-dabed5a1e064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.007593] env[70013]: DEBUG oslo_vmware.api [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230495, 'name': PowerOnVM_Task, 'duration_secs': 0.642804} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.007804] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 709.008026] env[70013]: INFO nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Took 10.72 seconds to spawn the instance on the hypervisor. [ 709.008204] env[70013]: DEBUG nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 709.009396] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d28913-4ef9-48fa-aac5-74a51dc8a568 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.083296] env[70013]: DEBUG nova.compute.manager [None req-e1331814-2ed3-41ab-b43b-c15fb2ad49c5 tempest-ServerDiagnosticsTest-1778048200 tempest-ServerDiagnosticsTest-1778048200-project-admin] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 709.085048] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bd8392-acd3-4a12-b073-4c91975e66eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.094753] env[70013]: INFO nova.compute.manager [None req-e1331814-2ed3-41ab-b43b-c15fb2ad49c5 tempest-ServerDiagnosticsTest-1778048200 tempest-ServerDiagnosticsTest-1778048200-project-admin] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Retrieving diagnostics [ 709.095905] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f016c404-7e8d-4ad7-a6f4-a459b76dd0dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.103623] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230499, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.146211] env[70013]: DEBUG nova.compute.utils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 709.149631] env[70013]: DEBUG nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 709.170882] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230496, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699932} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.171218] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9c5f2422-820e-454b-98be-695e3ac6d1e3/9c5f2422-820e-454b-98be-695e3ac6d1e3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 709.171451] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.171691] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14fb5045-d9eb-4f92-9718-c25358a3b552 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.183806] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 709.183806] env[70013]: value = "task-4230500" [ 709.183806] env[70013]: _type = "Task" [ 709.183806] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.194250] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230500, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.258020] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.258503] env[70013]: DEBUG nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Instance network_info: |[{"id": "b402fce4-7392-4139-8de5-7620820e0814", "address": "fa:16:3e:44:ef:b6", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb402fce4-73", "ovs_interfaceid": "b402fce4-7392-4139-8de5-7620820e0814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 709.259075] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:ef:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b402fce4-7392-4139-8de5-7620820e0814', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 709.268565] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Creating folder: Project (4f00f00d83424b5b8abfd6b22d9e119b). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 709.268955] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0decfcbb-019a-45a1-bd13-d2dc9ea94526 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.283840] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Created folder: Project (4f00f00d83424b5b8abfd6b22d9e119b) in parent group-v836999. [ 709.284115] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Creating folder: Instances. Parent ref: group-v837015. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 709.284631] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-984a741c-53c1-428f-9ab0-86d0d5390619 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.299743] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Created folder: Instances in parent group-v837015. [ 709.300128] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 709.300442] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 709.300721] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-892746ba-adc7-4854-99f8-cc09bbd855d4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.328078] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 709.328078] env[70013]: value = "task-4230503" [ 709.328078] env[70013]: _type = "Task" [ 709.328078] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.340069] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230503, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.397623] env[70013]: DEBUG oslo_concurrency.lockutils [req-85d8aabe-12fd-412d-9cae-8b8d7fa7a601 req-797c354a-5702-4d82-8bc2-8711c95a2202 service nova] Releasing lock "refresh_cache-9c5f2422-820e-454b-98be-695e3ac6d1e3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.464411] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 709.464734] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.465327] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 709.465327] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.465947] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 709.466492] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 709.466560] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 709.466776] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 709.467153] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 709.467153] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 709.467253] env[70013]: DEBUG nova.virt.hardware [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 709.468343] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a29d669-1358-4134-99f1-00532523d493 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.479892] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f43e1e-fe04-46b5-b013-7bc504d09f6c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.502114] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 709.508267] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 709.508627] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 709.508860] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a9f53b0-f78a-4144-823d-4d512ebfa743 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.536553] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 709.536553] env[70013]: value = "task-4230504" [ 709.536553] env[70013]: _type = "Task" [ 709.536553] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.544889] env[70013]: INFO nova.compute.manager [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Took 15.64 seconds to build instance. [ 709.552519] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230504, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.602837] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230499, 'name': CreateVM_Task, 'duration_secs': 0.851807} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.603013] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 709.603721] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.604436] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.604436] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 709.604608] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-776811d7-042a-4d07-8718-c7ad4a7aac78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.610170] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 709.610170] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e849b8-0671-4f0e-c7ab-94ae7f4702fc" [ 709.610170] env[70013]: _type = "Task" [ 709.610170] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.624899] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e849b8-0671-4f0e-c7ab-94ae7f4702fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.654231] env[70013]: DEBUG nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 709.697068] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230500, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137019} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.700573] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 709.701829] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f048264-db85-4bf0-ad10-cba013194626 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.731940] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 9c5f2422-820e-454b-98be-695e3ac6d1e3/9c5f2422-820e-454b-98be-695e3ac6d1e3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 709.735557] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-251aa578-cf26-4bb8-bb1d-8b05c9ce6130 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.758302] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 709.758302] env[70013]: value = "task-4230505" [ 709.758302] env[70013]: _type = "Task" [ 709.758302] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.772335] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230505, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.860428] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230503, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.048307] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b115d16-86a0-44bb-a64e-339516104152 tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.148s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.049905] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 12.328s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.050119] env[70013]: INFO nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] During sync_power_state the instance has a pending task (networking). Skip. [ 710.050296] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.061760] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230504, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.089070] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acadf3a1-42ba-41c1-8beb-42badfd1f984 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.101020] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de01c70-e9d3-4da5-84a0-86d02c597d96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.161421] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4720787-6814-4464-8c57-c3739e6d768b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.165428] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.165677] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.169330] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e849b8-0671-4f0e-c7ab-94ae7f4702fc, 'name': SearchDatastore_Task, 'duration_secs': 0.021737} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.170625] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 710.172278] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.172278] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.172278] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.172278] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.172278] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b6a0d00-d676-44eb-b66e-248a25214888 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.179673] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931c341d-d3d0-4f9a-a447-41ddc1f9d8fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.196721] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 710.196939] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 710.198031] env[70013]: DEBUG nova.compute.provider_tree [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.202215] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b626290-6249-4cdf-b4ea-a1ea624c998f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.206313] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 710.206313] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d05633-b1a9-0d93-b7aa-61cfc4c1652f" [ 710.206313] env[70013]: _type = "Task" [ 710.206313] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.215590] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d05633-b1a9-0d93-b7aa-61cfc4c1652f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.255436] env[70013]: DEBUG nova.compute.manager [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Received event network-vif-plugged-21a0b03c-bc07-44c7-b9e4-503338c106cd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 710.255691] env[70013]: DEBUG oslo_concurrency.lockutils [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] Acquiring lock "417f8c74-81cc-47d6-80ca-b61d03ebec50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.255775] env[70013]: DEBUG oslo_concurrency.lockutils [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.255957] env[70013]: DEBUG oslo_concurrency.lockutils [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.256127] env[70013]: DEBUG nova.compute.manager [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] No waiting events found dispatching network-vif-plugged-21a0b03c-bc07-44c7-b9e4-503338c106cd {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 710.256335] env[70013]: WARNING nova.compute.manager [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Received unexpected event network-vif-plugged-21a0b03c-bc07-44c7-b9e4-503338c106cd for instance with vm_state building and task_state spawning. [ 710.256475] env[70013]: DEBUG nova.compute.manager [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Received event network-changed-21a0b03c-bc07-44c7-b9e4-503338c106cd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 710.257398] env[70013]: DEBUG nova.compute.manager [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Refreshing instance network info cache due to event network-changed-21a0b03c-bc07-44c7-b9e4-503338c106cd. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 710.257398] env[70013]: DEBUG oslo_concurrency.lockutils [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] Acquiring lock "refresh_cache-417f8c74-81cc-47d6-80ca-b61d03ebec50" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.257398] env[70013]: DEBUG oslo_concurrency.lockutils [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] Acquired lock "refresh_cache-417f8c74-81cc-47d6-80ca-b61d03ebec50" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.257398] env[70013]: DEBUG nova.network.neutron [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Refreshing network info cache for port 21a0b03c-bc07-44c7-b9e4-503338c106cd {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.269030] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.349675] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230503, 'name': CreateVM_Task, 'duration_secs': 0.842445} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.349675] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 710.350372] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.350372] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.350654] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 710.351168] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2487befe-4b8a-4c74-9f7d-5b7c8cb8c9f5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.356702] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 710.356702] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cda203-ddf8-2f92-2588-6ebb061cf1e1" [ 710.356702] env[70013]: _type = "Task" [ 710.356702] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.369715] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cda203-ddf8-2f92-2588-6ebb061cf1e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.552173] env[70013]: DEBUG nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 710.562219] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230504, 'name': CreateVM_Task, 'duration_secs': 0.644648} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.562689] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 710.569404] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.672610] env[70013]: DEBUG nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 710.705438] env[70013]: DEBUG nova.compute.manager [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Received event network-vif-plugged-b402fce4-7392-4139-8de5-7620820e0814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 710.706150] env[70013]: DEBUG oslo_concurrency.lockutils [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] Acquiring lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.706886] env[70013]: DEBUG oslo_concurrency.lockutils [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.707435] env[70013]: DEBUG oslo_concurrency.lockutils [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.708450] env[70013]: DEBUG nova.compute.manager [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] No waiting events found dispatching network-vif-plugged-b402fce4-7392-4139-8de5-7620820e0814 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 710.708742] env[70013]: WARNING nova.compute.manager [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Received unexpected event network-vif-plugged-b402fce4-7392-4139-8de5-7620820e0814 for instance with vm_state building and task_state spawning. [ 710.712027] env[70013]: DEBUG nova.compute.manager [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Received event network-changed-b402fce4-7392-4139-8de5-7620820e0814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 710.712027] env[70013]: DEBUG nova.compute.manager [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Refreshing instance network info cache due to event network-changed-b402fce4-7392-4139-8de5-7620820e0814. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 710.712027] env[70013]: DEBUG oslo_concurrency.lockutils [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] Acquiring lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.712027] env[70013]: DEBUG oslo_concurrency.lockutils [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] Acquired lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.712027] env[70013]: DEBUG nova.network.neutron [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Refreshing network info cache for port b402fce4-7392-4139-8de5-7620820e0814 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.713489] env[70013]: DEBUG nova.scheduler.client.report [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 710.720880] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 710.722106] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.722106] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 710.722106] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.722106] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 710.722106] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 710.722601] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 710.722882] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 710.723188] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 710.723488] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 710.723778] env[70013]: DEBUG nova.virt.hardware [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 710.726016] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e88cee-7aa4-4a84-89f4-72a3529fec20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.750941] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d05633-b1a9-0d93-b7aa-61cfc4c1652f, 'name': SearchDatastore_Task, 'duration_secs': 0.047496} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.752373] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203d5173-f447-46af-a3a3-520fe51fa57f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.757613] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04599957-c704-423e-b357-fba389cbe05e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.778317] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 710.784077] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Creating folder: Project (50ebe9f1c15b4d0eabfad2efefcf6d37). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 710.786298] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3567bd3d-bf14-4491-85bf-7fa6a4b5662e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.788321] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 710.788321] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1896b-7d0b-c1cf-035e-3633b12dd1be" [ 710.788321] env[70013]: _type = "Task" [ 710.788321] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.791473] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230505, 'name': ReconfigVM_Task, 'duration_secs': 0.694914} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.795303] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 9c5f2422-820e-454b-98be-695e3ac6d1e3/9c5f2422-820e-454b-98be-695e3ac6d1e3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.796282] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d67d110e-ee03-439b-9fbf-c14cb5b6e085 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.803929] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 710.803929] env[70013]: value = "task-4230507" [ 710.803929] env[70013]: _type = "Task" [ 710.803929] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.809514] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1896b-7d0b-c1cf-035e-3633b12dd1be, 'name': SearchDatastore_Task, 'duration_secs': 0.031795} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.809768] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Created folder: Project (50ebe9f1c15b4d0eabfad2efefcf6d37) in parent group-v836999. [ 710.809931] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Creating folder: Instances. Parent ref: group-v837019. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 710.813599] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 710.815167] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 417f8c74-81cc-47d6-80ca-b61d03ebec50/417f8c74-81cc-47d6-80ca-b61d03ebec50.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 710.815167] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25761fd6-046b-4a83-adf8-322ba997b317 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.816166] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01b5f4ee-9ce4-41b2-923e-087d4c4506df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.829231] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230507, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.830942] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 710.830942] env[70013]: value = "task-4230508" [ 710.830942] env[70013]: _type = "Task" [ 710.830942] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.836589] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Created folder: Instances in parent group-v837019. [ 710.836838] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 710.840526] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 710.840908] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.841138] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb99dabf-b81f-4696-a488-e8bb01447fcd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.862087] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 710.862087] env[70013]: value = "task-4230510" [ 710.862087] env[70013]: _type = "Task" [ 710.862087] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.872467] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cda203-ddf8-2f92-2588-6ebb061cf1e1, 'name': SearchDatastore_Task, 'duration_secs': 0.014171} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.873901] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 710.874163] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.874414] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.874551] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.874724] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.874964] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230510, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.875223] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.875552] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 710.875780] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8ade958-e164-4b23-8384-c242aa597b82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.877780] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a56999aa-4687-40d7-8d36-b14d1809e69b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.884441] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 710.884441] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5222caa7-494b-0b91-ba36-c3b211276f0d" [ 710.884441] env[70013]: _type = "Task" [ 710.884441] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.889400] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 710.889574] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 710.890795] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fde2071e-fa11-49e7-80d2-26696cfd3b28 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.897475] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5222caa7-494b-0b91-ba36-c3b211276f0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.899036] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 710.899036] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd01f2-e4e4-fcb1-8f15-0cb1c5c49536" [ 710.899036] env[70013]: _type = "Task" [ 710.899036] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.908542] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd01f2-e4e4-fcb1-8f15-0cb1c5c49536, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.084097] env[70013]: DEBUG nova.network.neutron [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Successfully updated port: c9bdfb0a-f6f0-4467-b641-a8c04598910f {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 711.094438] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.231186] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.233497] env[70013]: DEBUG nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 711.236300] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.006s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.236407] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.236484] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 711.237526] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.771s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.238776] env[70013]: INFO nova.compute.claims [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.245035] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8fff43-68f6-45d0-84cb-1524f9edd610 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.256913] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71f1012-f6c1-4e07-b7cb-aad81674884f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.278657] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2640b7c-6877-411c-a676-8088995a408e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.291456] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b970fc-1462-4d72-bdf5-6582c6126efd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.330354] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180791MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 711.330534] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.356829] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230508, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.361660] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230507, 'name': Rename_Task, 'duration_secs': 0.182571} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.366478] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 711.367207] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0f74a2f-ce26-4374-bd60-b59a621663ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.390514] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230510, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.398824] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 711.398824] env[70013]: value = "task-4230511" [ 711.398824] env[70013]: _type = "Task" [ 711.398824] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.411080] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5222caa7-494b-0b91-ba36-c3b211276f0d, 'name': SearchDatastore_Task, 'duration_secs': 0.022267} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.411432] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.411708] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 711.411889] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.416607] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230511, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.421170] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd01f2-e4e4-fcb1-8f15-0cb1c5c49536, 'name': SearchDatastore_Task, 'duration_secs': 0.010915} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.422407] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1536b87d-c670-4b9f-87c4-dad2a1b26c23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.429376] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 711.429376] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca489f-9cc5-afd5-a6cc-7b86913ce4a6" [ 711.429376] env[70013]: _type = "Task" [ 711.429376] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.439516] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca489f-9cc5-afd5-a6cc-7b86913ce4a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.445213] env[70013]: DEBUG nova.network.neutron [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Updated VIF entry in instance network info cache for port 21a0b03c-bc07-44c7-b9e4-503338c106cd. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 711.446510] env[70013]: DEBUG nova.network.neutron [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Updating instance_info_cache with network_info: [{"id": "21a0b03c-bc07-44c7-b9e4-503338c106cd", "address": "fa:16:3e:8e:9d:ae", "network": {"id": "8aa548f3-6ae4-440d-8d55-f22e41739121", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1100728332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8214085b60d0482596cdb3e8df48c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21a0b03c-bc", "ovs_interfaceid": "21a0b03c-bc07-44c7-b9e4-503338c106cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.589504] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "refresh_cache-9a553d85-5086-46c5-8df4-451928e38ed9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.589729] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquired lock "refresh_cache-9a553d85-5086-46c5-8df4-451928e38ed9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.589914] env[70013]: DEBUG nova.network.neutron [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.632424] env[70013]: DEBUG nova.network.neutron [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updated VIF entry in instance network info cache for port b402fce4-7392-4139-8de5-7620820e0814. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 711.632748] env[70013]: DEBUG nova.network.neutron [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance_info_cache with network_info: [{"id": "b402fce4-7392-4139-8de5-7620820e0814", "address": "fa:16:3e:44:ef:b6", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb402fce4-73", "ovs_interfaceid": "b402fce4-7392-4139-8de5-7620820e0814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.744580] env[70013]: DEBUG nova.compute.utils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 711.746594] env[70013]: DEBUG nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 711.746934] env[70013]: DEBUG nova.network.neutron [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 711.795109] env[70013]: DEBUG nova.policy [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d6c00fc04b8435388ae5e2652f8e0eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51b2e9e21dcc447d86aa3d2f5f307d4f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 711.848232] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.848232] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.848232] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.848232] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.848417] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.852891] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613941} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.853400] env[70013]: INFO nova.compute.manager [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Terminating instance [ 711.859188] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 417f8c74-81cc-47d6-80ca-b61d03ebec50/417f8c74-81cc-47d6-80ca-b61d03ebec50.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 711.859188] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 711.860509] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-67b8a13a-a9d5-46b8-87cd-ffd2e2cf0cf7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.872711] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 711.872711] env[70013]: value = "task-4230512" [ 711.872711] env[70013]: _type = "Task" [ 711.872711] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.880650] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230510, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.886999] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230512, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.909430] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230511, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.941181] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca489f-9cc5-afd5-a6cc-7b86913ce4a6, 'name': SearchDatastore_Task, 'duration_secs': 0.013614} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.941468] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.941788] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 3b29c99a-0d55-40b1-a155-199b1ced2146/3b29c99a-0d55-40b1-a155-199b1ced2146.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 711.942140] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.942247] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 711.942481] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b1f345a-9758-4814-822e-0de46b601c74 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.944547] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c861c442-c473-4caa-8953-d532452f77f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.948549] env[70013]: DEBUG oslo_concurrency.lockutils [req-beb87c8f-7da0-468c-926c-f4c44297250f req-9a66fd60-ac16-4868-957e-4afc500c5f4e service nova] Releasing lock "refresh_cache-417f8c74-81cc-47d6-80ca-b61d03ebec50" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.953551] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 711.953551] env[70013]: value = "task-4230513" [ 711.953551] env[70013]: _type = "Task" [ 711.953551] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.958932] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 711.959214] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 711.961059] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fff34643-abc2-4b27-ab6e-0d2b0156a90d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.966847] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230513, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.971223] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 711.971223] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521566c3-fa0a-4470-eef1-4fcde546658f" [ 711.971223] env[70013]: _type = "Task" [ 711.971223] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.981267] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521566c3-fa0a-4470-eef1-4fcde546658f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.136416] env[70013]: DEBUG oslo_concurrency.lockutils [req-87ca25f7-c334-49a1-99f2-40a25505a40d req-e9347cad-6914-4f92-b8b5-b18e8da1050a service nova] Releasing lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 712.180802] env[70013]: DEBUG nova.network.neutron [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.257270] env[70013]: DEBUG nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 712.271837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Acquiring lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.271837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.271837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Acquiring lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.271837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.272095] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.272989] env[70013]: INFO nova.compute.manager [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Terminating instance [ 712.365942] env[70013]: DEBUG nova.compute.manager [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 712.366645] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.374778] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9a9b97-2bfd-4f26-a4cc-026e9f3c9ffb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.396935] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 712.404238] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-891deb55-534b-485c-b607-ac2628e35809 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.406967] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230512, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.219189} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.406967] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230510, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.410714] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 712.418156] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcc20da-203a-4217-9e34-8542e7de5920 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.426192] env[70013]: DEBUG oslo_vmware.api [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 712.426192] env[70013]: value = "task-4230514" [ 712.426192] env[70013]: _type = "Task" [ 712.426192] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.450186] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] 417f8c74-81cc-47d6-80ca-b61d03ebec50/417f8c74-81cc-47d6-80ca-b61d03ebec50.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 712.450586] env[70013]: DEBUG oslo_vmware.api [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230511, 'name': PowerOnVM_Task, 'duration_secs': 1.018078} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.454574] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27393aab-6241-4407-b323-8efeb608577c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.469673] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 712.469933] env[70013]: INFO nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Took 11.89 seconds to spawn the instance on the hypervisor. [ 712.470149] env[70013]: DEBUG nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 712.477853] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4386b4-98d4-4cf1-9961-0092b4662dd6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.481262] env[70013]: DEBUG oslo_vmware.api [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.496336] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 712.496336] env[70013]: value = "task-4230515" [ 712.496336] env[70013]: _type = "Task" [ 712.496336] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.517640] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230513, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.526163] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521566c3-fa0a-4470-eef1-4fcde546658f, 'name': SearchDatastore_Task, 'duration_secs': 0.031518} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.532384] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230515, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.532384] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64392d84-0307-490e-815d-4579290a968d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.541654] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 712.541654] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522c3e2c-007b-3199-ea28-f1c719c464d2" [ 712.541654] env[70013]: _type = "Task" [ 712.541654] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.557023] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522c3e2c-007b-3199-ea28-f1c719c464d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.558570] env[70013]: DEBUG nova.network.neutron [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Successfully created port: a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.642858] env[70013]: DEBUG nova.network.neutron [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Updating instance_info_cache with network_info: [{"id": "c9bdfb0a-f6f0-4467-b641-a8c04598910f", "address": "fa:16:3e:46:3c:57", "network": {"id": "1624bc7c-733b-48df-9707-566347532dac", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-739181955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b4f41a91a5d4b679adafa6624a1db74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9bdfb0a-f6", "ovs_interfaceid": "c9bdfb0a-f6f0-4467-b641-a8c04598910f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.704669] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9080a7f4-1e14-4ca4-8efd-8b7a90da16c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.715680] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7a854d-e19d-43b4-9422-cae27663fc3b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.768953] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbae25f-a429-4e8c-a6ed-830a4379ec7b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.774894] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ae7f8d-25ec-4b71-8ba1-1718cd49623a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.783118] env[70013]: DEBUG nova.compute.manager [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 712.784665] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.795186] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04ff742-144c-4ca5-8913-2e214cfe3b03 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.801128] env[70013]: DEBUG nova.compute.provider_tree [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.808218] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 712.808839] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23b5a955-2e3f-4472-8192-8925bdd8b01f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.816343] env[70013]: DEBUG oslo_vmware.api [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Waiting for the task: (returnval){ [ 712.816343] env[70013]: value = "task-4230516" [ 712.816343] env[70013]: _type = "Task" [ 712.816343] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.828417] env[70013]: DEBUG oslo_vmware.api [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Task: {'id': task-4230516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.893719] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230510, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.934668] env[70013]: DEBUG oslo_vmware.api [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230514, 'name': PowerOffVM_Task, 'duration_secs': 0.26899} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.936040] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 712.936040] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 712.936040] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a51ff11-ae46-442d-97bf-910473a4ef43 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.991267] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230513, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.839156} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.991267] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 3b29c99a-0d55-40b1-a155-199b1ced2146/3b29c99a-0d55-40b1-a155-199b1ced2146.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 712.992637] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.992637] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29b0393a-c6fc-4a4f-aad1-b82ef44f55fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.000356] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 713.000356] env[70013]: value = "task-4230518" [ 713.000356] env[70013]: _type = "Task" [ 713.000356] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.013991] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 713.013991] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 713.014259] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Deleting the datastore file [datastore2] 2b91ec52-2d32-4f86-acf2-f74f661f65b5 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 713.014968] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ba43ab7-6f93-42dd-ac3e-402a15d9174b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.027144] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230515, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.032108] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.033164] env[70013]: INFO nova.compute.manager [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Took 18.56 seconds to build instance. [ 713.038507] env[70013]: DEBUG oslo_vmware.api [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for the task: (returnval){ [ 713.038507] env[70013]: value = "task-4230519" [ 713.038507] env[70013]: _type = "Task" [ 713.038507] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.056241] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522c3e2c-007b-3199-ea28-f1c719c464d2, 'name': SearchDatastore_Task, 'duration_secs': 0.062102} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.060853] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 713.061221] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 713.061636] env[70013]: DEBUG oslo_vmware.api [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230519, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.062233] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63c5145a-8f20-47dd-b3a5-12e4ca097ee5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.074789] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 713.074789] env[70013]: value = "task-4230520" [ 713.074789] env[70013]: _type = "Task" [ 713.074789] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.086191] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230520, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.100053] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "f12b8575-5082-4be9-9bf5-f4279860d19d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 713.100636] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.144298] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Releasing lock "refresh_cache-9a553d85-5086-46c5-8df4-451928e38ed9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 713.144298] env[70013]: DEBUG nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Instance network_info: |[{"id": "c9bdfb0a-f6f0-4467-b641-a8c04598910f", "address": "fa:16:3e:46:3c:57", "network": {"id": "1624bc7c-733b-48df-9707-566347532dac", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-739181955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b4f41a91a5d4b679adafa6624a1db74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9bdfb0a-f6", "ovs_interfaceid": "c9bdfb0a-f6f0-4467-b641-a8c04598910f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 713.144876] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:3c:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c9bdfb0a-f6f0-4467-b641-a8c04598910f', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.153042] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Creating folder: Project (1b4f41a91a5d4b679adafa6624a1db74). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 713.153286] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d11a4294-fc3d-4622-ba40-ea171301bfcb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.168726] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Created folder: Project (1b4f41a91a5d4b679adafa6624a1db74) in parent group-v836999. [ 713.168726] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Creating folder: Instances. Parent ref: group-v837022. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 713.168726] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03a7a8e6-1fa2-44e7-848c-f89b79f2b076 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.183865] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Created folder: Instances in parent group-v837022. [ 713.184163] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 713.184404] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 713.184630] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1c8bcc7-eefa-4acd-9d6f-8d7ea50563e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.211939] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.211939] env[70013]: value = "task-4230523" [ 713.211939] env[70013]: _type = "Task" [ 713.211939] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.219886] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230523, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.273142] env[70013]: DEBUG nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 713.303792] env[70013]: DEBUG nova.scheduler.client.report [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 713.314435] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 713.314682] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.314833] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 713.315029] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.315192] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 713.315350] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 713.315560] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 713.315717] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 713.315935] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 713.316153] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 713.316343] env[70013]: DEBUG nova.virt.hardware [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 713.317386] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbd5219-10c3-4185-a1d2-68f3cd633c23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.340378] env[70013]: DEBUG oslo_vmware.api [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Task: {'id': task-4230516, 'name': PowerOffVM_Task, 'duration_secs': 0.349852} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.341948] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 713.341948] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 713.343551] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920d6dd2-ab91-4eb1-8b96-a931080c2b3f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.349568] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48e9c1e7-9eed-45ac-ada9-402a542c4b7e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.390316] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230510, 'name': CreateVM_Task, 'duration_secs': 2.48144} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.390443] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 713.390882] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.391126] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.391552] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 713.391796] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7c8f5e1-5166-4201-830c-09675e751a4d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.399574] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 713.399574] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5223d603-b2a6-6220-0cc7-910eadf339b8" [ 713.399574] env[70013]: _type = "Task" [ 713.399574] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.411658] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5223d603-b2a6-6220-0cc7-910eadf339b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.435233] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 713.435366] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 713.435493] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Deleting the datastore file [datastore2] e8998cab-533b-46b1-a1ea-5eb5db7d6d99 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 713.436146] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-726a38e6-493a-47be-a721-01c8f03018bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.445689] env[70013]: DEBUG oslo_vmware.api [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Waiting for the task: (returnval){ [ 713.445689] env[70013]: value = "task-4230525" [ 713.445689] env[70013]: _type = "Task" [ 713.445689] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.456976] env[70013]: DEBUG oslo_vmware.api [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Task: {'id': task-4230525, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.507870] env[70013]: DEBUG nova.compute.manager [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Received event network-vif-plugged-c9bdfb0a-f6f0-4467-b641-a8c04598910f {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 713.508010] env[70013]: DEBUG oslo_concurrency.lockutils [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] Acquiring lock "9a553d85-5086-46c5-8df4-451928e38ed9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 713.508329] env[70013]: DEBUG oslo_concurrency.lockutils [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] Lock "9a553d85-5086-46c5-8df4-451928e38ed9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.508797] env[70013]: DEBUG oslo_concurrency.lockutils [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] Lock "9a553d85-5086-46c5-8df4-451928e38ed9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.508797] env[70013]: DEBUG nova.compute.manager [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] No waiting events found dispatching network-vif-plugged-c9bdfb0a-f6f0-4467-b641-a8c04598910f {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 713.508797] env[70013]: WARNING nova.compute.manager [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Received unexpected event network-vif-plugged-c9bdfb0a-f6f0-4467-b641-a8c04598910f for instance with vm_state building and task_state spawning. [ 713.508950] env[70013]: DEBUG nova.compute.manager [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Received event network-changed-c9bdfb0a-f6f0-4467-b641-a8c04598910f {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 713.509111] env[70013]: DEBUG nova.compute.manager [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Refreshing instance network info cache due to event network-changed-c9bdfb0a-f6f0-4467-b641-a8c04598910f. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 713.509306] env[70013]: DEBUG oslo_concurrency.lockutils [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] Acquiring lock "refresh_cache-9a553d85-5086-46c5-8df4-451928e38ed9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.509493] env[70013]: DEBUG oslo_concurrency.lockutils [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] Acquired lock "refresh_cache-9a553d85-5086-46c5-8df4-451928e38ed9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.509685] env[70013]: DEBUG nova.network.neutron [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Refreshing network info cache for port c9bdfb0a-f6f0-4467-b641-a8c04598910f {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.519422] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230515, 'name': ReconfigVM_Task, 'duration_secs': 0.694657} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.520119] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Reconfigured VM instance instance-00000005 to attach disk [datastore1] 417f8c74-81cc-47d6-80ca-b61d03ebec50/417f8c74-81cc-47d6-80ca-b61d03ebec50.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 713.521360] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3391e59b-8dde-4103-8b9d-917283ec6341 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.526646] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081116} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.527356] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 713.528198] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31110134-bff4-4816-8b66-6b438d7bcc5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.536878] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5dd4d85-b39d-4fbd-8b3c-c22453db4ea4 tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.077s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.538077] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.815s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.538077] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 713.538077] env[70013]: value = "task-4230526" [ 713.538077] env[70013]: _type = "Task" [ 713.538077] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.550177] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc7d80f-86bc-4beb-9e0a-fbe43419b115 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.562885] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] 3b29c99a-0d55-40b1-a155-199b1ced2146/3b29c99a-0d55-40b1-a155-199b1ced2146.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 713.573391] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1851f2b7-134a-4a8c-96ee-826e8f96f84a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.604455] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230526, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.606051] env[70013]: DEBUG oslo_vmware.api [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Task: {'id': task-4230519, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230681} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.609321] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 713.609567] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 713.609747] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.609923] env[70013]: INFO nova.compute.manager [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Took 1.24 seconds to destroy the instance on the hypervisor. [ 713.610191] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 713.610457] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 713.610457] env[70013]: value = "task-4230527" [ 713.610457] env[70013]: _type = "Task" [ 713.610457] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.610988] env[70013]: DEBUG nova.compute.manager [-] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 713.611098] env[70013]: DEBUG nova.network.neutron [-] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.625208] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.628586] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230520, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.721745] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230523, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.829073] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.829073] env[70013]: DEBUG nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 713.831790] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.291s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.833275] env[70013]: INFO nova.compute.claims [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.912223] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5223d603-b2a6-6220-0cc7-910eadf339b8, 'name': SearchDatastore_Task, 'duration_secs': 0.051858} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.912556] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 713.912866] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 713.913142] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.913362] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.913584] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 713.914268] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80c3d582-41b0-411c-966a-88ba92e951ee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.925551] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 713.925782] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 713.927256] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fb7edb8-87aa-469b-a110-b1047773118f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.934197] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 713.934197] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b38cbc-5749-4262-cbf6-6d511f943298" [ 713.934197] env[70013]: _type = "Task" [ 713.934197] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.947446] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b38cbc-5749-4262-cbf6-6d511f943298, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.957766] env[70013]: DEBUG oslo_vmware.api [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Task: {'id': task-4230525, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.404134} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.958075] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 713.958798] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 713.958798] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.958798] env[70013]: INFO nova.compute.manager [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Took 1.18 seconds to destroy the instance on the hypervisor. [ 713.959022] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 713.959382] env[70013]: DEBUG nova.compute.manager [-] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 713.959382] env[70013]: DEBUG nova.network.neutron [-] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.048918] env[70013]: DEBUG nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 714.063519] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230526, 'name': Rename_Task, 'duration_secs': 0.318544} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.067598] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 714.068562] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2e39084-f574-48a2-99c2-ae3d97e13ee0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.076714] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 714.076714] env[70013]: value = "task-4230528" [ 714.076714] env[70013]: _type = "Task" [ 714.076714] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.091023] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230528, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.103147] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230520, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.630479} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.103599] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 714.104154] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 714.104785] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cffa65fc-33e9-4c8c-895b-22cfbb105cea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.110810] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.574s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.115511] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 714.115511] env[70013]: value = "task-4230529" [ 714.115511] env[70013]: _type = "Task" [ 714.115511] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.135587] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.139332] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230527, 'name': ReconfigVM_Task, 'duration_secs': 0.399594} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.139591] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Reconfigured VM instance instance-00000006 to attach disk [datastore1] 3b29c99a-0d55-40b1-a155-199b1ced2146/3b29c99a-0d55-40b1-a155-199b1ced2146.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 714.140224] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1f64526-388a-49df-9da5-d68a2e84a99b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.146884] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 714.146884] env[70013]: value = "task-4230530" [ 714.146884] env[70013]: _type = "Task" [ 714.146884] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.161741] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230530, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.224238] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230523, 'name': CreateVM_Task, 'duration_secs': 0.617108} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.224397] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 714.225057] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.225355] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.225573] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 714.225937] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0857cdc3-11a0-486b-80c8-8070bd1aca50 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.233778] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 714.233778] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5273672d-de1d-be6f-e41b-5b95d5df79e9" [ 714.233778] env[70013]: _type = "Task" [ 714.233778] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.242073] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5273672d-de1d-be6f-e41b-5b95d5df79e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.340564] env[70013]: DEBUG nova.compute.utils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 714.344843] env[70013]: DEBUG nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 714.345211] env[70013]: DEBUG nova.network.neutron [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 714.446231] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b38cbc-5749-4262-cbf6-6d511f943298, 'name': SearchDatastore_Task, 'duration_secs': 0.054389} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.446974] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67816059-8c06-4062-9ac2-3f22895eb205 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.453185] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 714.453185] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b57090-2592-f27f-5661-00997c7cb752" [ 714.453185] env[70013]: _type = "Task" [ 714.453185] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.463334] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b57090-2592-f27f-5661-00997c7cb752, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.550972] env[70013]: DEBUG nova.network.neutron [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Updated VIF entry in instance network info cache for port c9bdfb0a-f6f0-4467-b641-a8c04598910f. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 714.551349] env[70013]: DEBUG nova.network.neutron [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Updating instance_info_cache with network_info: [{"id": "c9bdfb0a-f6f0-4467-b641-a8c04598910f", "address": "fa:16:3e:46:3c:57", "network": {"id": "1624bc7c-733b-48df-9707-566347532dac", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-739181955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b4f41a91a5d4b679adafa6624a1db74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9bdfb0a-f6", "ovs_interfaceid": "c9bdfb0a-f6f0-4467-b641-a8c04598910f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.581656] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.587157] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230528, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.588765] env[70013]: DEBUG nova.policy [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96992dbd2bd34620aa97d5e3931c73ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e94ebd7a191d4701a62223fbc2e1bdf8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 714.632925] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.162442} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.633227] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.634021] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b97972-109c-4277-90b8-2fdea7f71c08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.657665] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.661772] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c18e1ed-d36f-4ac0-85a9-b997babf57d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.683076] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230530, 'name': Rename_Task, 'duration_secs': 0.282113} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.684591] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 714.684999] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 714.684999] env[70013]: value = "task-4230531" [ 714.684999] env[70013]: _type = "Task" [ 714.684999] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.685252] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14ccb211-a003-4008-81ca-aae830da403d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.700466] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 714.700466] env[70013]: value = "task-4230532" [ 714.700466] env[70013]: _type = "Task" [ 714.700466] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.709685] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.752033] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5273672d-de1d-be6f-e41b-5b95d5df79e9, 'name': SearchDatastore_Task, 'duration_secs': 0.03105} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.752135] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.752333] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 714.752562] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.846501] env[70013]: DEBUG nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 714.970073] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b57090-2592-f27f-5661-00997c7cb752, 'name': SearchDatastore_Task, 'duration_secs': 0.04964} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.970287] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.970586] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 714.970918] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.971191] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 714.971392] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-239b8639-5752-416d-876c-1d573d3eb47b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.977444] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8d9d67a-a224-48fd-bc41-2767a9001607 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.980536] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.981358] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.981358] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "9c5f2422-820e-454b-98be-695e3ac6d1e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.981358] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.981707] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.986290] env[70013]: INFO nova.compute.manager [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Terminating instance [ 714.993582] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 714.993582] env[70013]: value = "task-4230533" [ 714.993582] env[70013]: _type = "Task" [ 714.993582] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.001400] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.001400] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 715.006966] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06ec8d94-30e4-4db7-92a5-5a6be43aa120 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.016785] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.021961] env[70013]: DEBUG nova.network.neutron [-] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.023430] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 715.023430] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b12167-2bcf-b17c-b706-c103c25cb2af" [ 715.023430] env[70013]: _type = "Task" [ 715.023430] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.040460] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b12167-2bcf-b17c-b706-c103c25cb2af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.057554] env[70013]: DEBUG oslo_concurrency.lockutils [req-965e3b9e-7f4b-4e3c-b6f7-5f4359094bcb req-17b08aaf-03e4-4f45-ab1a-0df47a0511b5 service nova] Releasing lock "refresh_cache-9a553d85-5086-46c5-8df4-451928e38ed9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.064533] env[70013]: DEBUG nova.network.neutron [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Successfully created port: aa55189e-9d4e-4fc1-9143-9b0ade9887b7 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.092227] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230528, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.113088] env[70013]: DEBUG nova.network.neutron [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Successfully updated port: a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 715.115964] env[70013]: DEBUG nova.network.neutron [-] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.200889] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.214214] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230532, 'name': PowerOnVM_Task} progress is 90%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.215221] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc4484a-44a4-494c-adec-5265fc2053d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.224433] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726cceb8-1fa3-47bc-bcfc-e615e4f2b557 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.260937] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5051a5f-3817-4f72-9349-548d1984ae3c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.270303] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b8799c-54f2-4dce-9a38-237c64686bea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.286507] env[70013]: DEBUG nova.compute.provider_tree [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.492922] env[70013]: DEBUG nova.compute.manager [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 715.492922] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.492922] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8bd207-6e91-4e59-9191-dbaad976084a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.517055] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230533, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.517550] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 715.517982] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffe2a81c-624c-491e-b0e2-5f9aed4e93b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.525302] env[70013]: INFO nova.compute.manager [-] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Took 1.57 seconds to deallocate network for instance. [ 715.534227] env[70013]: DEBUG oslo_vmware.api [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 715.534227] env[70013]: value = "task-4230534" [ 715.534227] env[70013]: _type = "Task" [ 715.534227] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.551344] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b12167-2bcf-b17c-b706-c103c25cb2af, 'name': SearchDatastore_Task, 'duration_secs': 0.021754} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.554574] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-934e1b47-d2b6-462c-8bce-05428cff5b45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.562039] env[70013]: DEBUG oslo_vmware.api [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.566771] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 715.566771] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5247e2ca-f4cb-45c6-18c5-f32dcad6adf4" [ 715.566771] env[70013]: _type = "Task" [ 715.566771] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.576744] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5247e2ca-f4cb-45c6-18c5-f32dcad6adf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.587861] env[70013]: DEBUG oslo_vmware.api [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230528, 'name': PowerOnVM_Task, 'duration_secs': 1.029333} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.587961] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 715.588183] env[70013]: INFO nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Took 12.59 seconds to spawn the instance on the hypervisor. [ 715.588902] env[70013]: DEBUG nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 715.589887] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd017a7-11a5-44ca-8a5e-6c4e8a4346a2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.617461] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.617562] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 715.617826] env[70013]: DEBUG nova.network.neutron [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.619673] env[70013]: INFO nova.compute.manager [-] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Took 2.01 seconds to deallocate network for instance. [ 715.704283] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230531, 'name': ReconfigVM_Task, 'duration_secs': 0.572705} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.707028] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.707981] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c585092-d48e-4e39-81a2-8b6620b31347 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.717049] env[70013]: DEBUG oslo_vmware.api [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230532, 'name': PowerOnVM_Task, 'duration_secs': 0.849308} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.719132] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 715.719492] env[70013]: INFO nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Took 10.15 seconds to spawn the instance on the hypervisor. [ 715.720542] env[70013]: DEBUG nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 715.721238] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 715.721238] env[70013]: value = "task-4230535" [ 715.721238] env[70013]: _type = "Task" [ 715.721238] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.722701] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7450f8-7cd0-4e8b-a771-88955615e612 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.743559] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230535, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.789754] env[70013]: DEBUG nova.scheduler.client.report [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 715.869379] env[70013]: DEBUG nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 715.906817] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 715.907312] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.907596] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 715.907884] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.908189] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 715.908477] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 715.908858] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 715.909517] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 715.909517] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 715.909644] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 715.909779] env[70013]: DEBUG nova.virt.hardware [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 715.911321] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f9569a-c56c-4bcb-b4f7-db77287b9735 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.924950] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7869aff1-fd89-4f83-b07d-662bfdd243fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.007172] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658157} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.007435] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 716.007649] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 716.007900] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-208f7591-332c-4da6-9f4d-ceaaae151919 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.017040] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 716.017040] env[70013]: value = "task-4230536" [ 716.017040] env[70013]: _type = "Task" [ 716.017040] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.025971] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230536, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.044421] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.056184] env[70013]: DEBUG oslo_vmware.api [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230534, 'name': PowerOffVM_Task, 'duration_secs': 0.516584} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.056184] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 716.056184] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 716.056184] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc8ea5b4-9c97-47df-9d83-c8a09484a341 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.071603] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.071773] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.082391] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5247e2ca-f4cb-45c6-18c5-f32dcad6adf4, 'name': SearchDatastore_Task, 'duration_secs': 0.044607} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.082463] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.084333] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9a553d85-5086-46c5-8df4-451928e38ed9/9a553d85-5086-46c5-8df4-451928e38ed9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 716.084333] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88c75dd6-cee7-413b-8f27-df72a98a78da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.093027] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 716.093027] env[70013]: value = "task-4230538" [ 716.093027] env[70013]: _type = "Task" [ 716.093027] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.110543] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.112900] env[70013]: INFO nova.compute.manager [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Took 20.78 seconds to build instance. [ 716.130991] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.132917] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 716.132917] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 716.133317] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Deleting the datastore file [datastore1] 9c5f2422-820e-454b-98be-695e3ac6d1e3 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 716.133762] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e35c8d5-b3ad-4197-befa-31a49a1b07c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.143345] env[70013]: DEBUG oslo_vmware.api [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for the task: (returnval){ [ 716.143345] env[70013]: value = "task-4230539" [ 716.143345] env[70013]: _type = "Task" [ 716.143345] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.154983] env[70013]: DEBUG oslo_vmware.api [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.215193] env[70013]: DEBUG nova.network.neutron [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.242251] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230535, 'name': Rename_Task, 'duration_secs': 0.348055} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.243756] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 716.243756] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-506859db-7dfb-4586-bb15-81eb506e3caf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.255458] env[70013]: INFO nova.compute.manager [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Took 20.14 seconds to build instance. [ 716.256500] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 716.256500] env[70013]: value = "task-4230540" [ 716.256500] env[70013]: _type = "Task" [ 716.256500] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.267761] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230540, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.296686] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 716.297440] env[70013]: DEBUG nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 716.302403] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.521s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.304032] env[70013]: INFO nova.compute.claims [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.444837] env[70013]: DEBUG nova.compute.manager [req-380d39cf-f8dc-4901-8563-6f4b584a9a6d req-1cee24c6-5ae4-4d7d-a460-5e73a32b7a19 service nova] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Received event network-vif-deleted-63c17eba-d2d2-44c2-bd66-bcbc6d27e157 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 716.528702] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230536, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.160809} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.528702] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 716.529520] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85533cec-1f0e-450d-99af-57872c3da7a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.554204] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 716.557092] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7f12bf8-b872-4d5f-a10c-41bafae9b40c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.580740] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 716.580740] env[70013]: value = "task-4230541" [ 716.580740] env[70013]: _type = "Task" [ 716.580740] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.590770] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230541, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.603543] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230538, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.615608] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3f668bbd-39cb-4f0c-aaa9-1c6f16b4b12a tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.297s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 716.655750] env[70013]: DEBUG oslo_vmware.api [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Task: {'id': task-4230539, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.372104} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.656083] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 716.656286] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 716.656505] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.656709] env[70013]: INFO nova.compute.manager [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Took 1.17 seconds to destroy the instance on the hypervisor. [ 716.656999] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 716.657255] env[70013]: DEBUG nova.compute.manager [-] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 716.657385] env[70013]: DEBUG nova.network.neutron [-] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.739732] env[70013]: DEBUG nova.network.neutron [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.758139] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c84fc8da-e6b7-48e6-ad27-76a91cd08613 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.659s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 716.773239] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230540, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.804531] env[70013]: DEBUG nova.compute.utils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 716.806207] env[70013]: DEBUG nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 716.806438] env[70013]: DEBUG nova.network.neutron [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.920701] env[70013]: DEBUG nova.policy [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd357d73286774d2cbe0ab68494c7c57b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75ec7d6e9ea54c06bd3b7605f16b391a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 717.080115] env[70013]: DEBUG nova.compute.manager [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Received event network-vif-deleted-ecf33e4a-0e1b-41fe-b723-bd679a50d3a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 717.080115] env[70013]: DEBUG nova.compute.manager [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Received event network-vif-plugged-a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 717.080115] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.080115] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] Lock "28a56a75-9b30-4121-8252-a9e57287441c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.085027] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] Lock "28a56a75-9b30-4121-8252-a9e57287441c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.085027] env[70013]: DEBUG nova.compute.manager [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] No waiting events found dispatching network-vif-plugged-a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 717.085027] env[70013]: WARNING nova.compute.manager [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Received unexpected event network-vif-plugged-a0d11944-d177-42a5-ad2b-22c116396f8a for instance with vm_state building and task_state spawning. [ 717.085027] env[70013]: DEBUG nova.compute.manager [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Received event network-changed-a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 717.085027] env[70013]: DEBUG nova.compute.manager [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Refreshing instance network info cache due to event network-changed-a0d11944-d177-42a5-ad2b-22c116396f8a. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 717.085578] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.098018] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.106668] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637948} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.107713] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9a553d85-5086-46c5-8df4-451928e38ed9/9a553d85-5086-46c5-8df4-451928e38ed9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 717.108123] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.108559] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34fee46b-15d6-4695-aa33-67629ee8e3e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.119021] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 717.119021] env[70013]: value = "task-4230542" [ 717.119021] env[70013]: _type = "Task" [ 717.119021] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.122049] env[70013]: DEBUG nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 717.134571] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.244430] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 717.244931] env[70013]: DEBUG nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Instance network_info: |[{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 717.245422] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.245587] env[70013]: DEBUG nova.network.neutron [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Refreshing network info cache for port a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.248107] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:10:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'abcf0d10-3f3f-45dc-923e-1c78766e2dad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0d11944-d177-42a5-ad2b-22c116396f8a', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.257390] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Creating folder: Project (51b2e9e21dcc447d86aa3d2f5f307d4f). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.261284] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-067fa1ba-b686-4fad-a882-288a2b93bcd0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.269921] env[70013]: DEBUG nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 717.284681] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230540, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.284681] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Created folder: Project (51b2e9e21dcc447d86aa3d2f5f307d4f) in parent group-v836999. [ 717.284681] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Creating folder: Instances. Parent ref: group-v837025. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.284681] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-38f57501-b894-4675-8921-2bfaac04cbdf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.298304] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Created folder: Instances in parent group-v837025. [ 717.298585] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 717.298776] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 717.299258] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-121772e7-152e-4ae5-85c5-5fe6164b0a69 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.320195] env[70013]: DEBUG nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 717.333694] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.333694] env[70013]: value = "task-4230545" [ 717.333694] env[70013]: _type = "Task" [ 717.333694] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.344787] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230545, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.569860] env[70013]: DEBUG nova.network.neutron [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Successfully updated port: aa55189e-9d4e-4fc1-9143-9b0ade9887b7 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 717.600368] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230541, 'name': ReconfigVM_Task, 'duration_secs': 0.840027} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.600644] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Reconfigured VM instance instance-00000008 to attach disk [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 717.601310] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3856f22-b474-4942-98e6-5264d97302bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.609993] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 717.609993] env[70013]: value = "task-4230546" [ 717.609993] env[70013]: _type = "Task" [ 717.609993] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.618489] env[70013]: DEBUG nova.network.neutron [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Successfully created port: a4ff3dbe-e842-4e4a-939e-e3f429ebad04 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.630088] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230546, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.642037] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078626} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.643720] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 717.645499] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17324a03-8639-4a57-b246-5e1b431f64c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.656170] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.819234] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] 9a553d85-5086-46c5-8df4-451928e38ed9/9a553d85-5086-46c5-8df4-451928e38ed9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 717.819234] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60a16ab5-6235-4774-aa6e-a4b6117ff88e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.819234] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 717.819234] env[70013]: value = "task-4230547" [ 717.819234] env[70013]: _type = "Task" [ 717.819234] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.819234] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230547, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.819234] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6232a74-229e-4baa-8afb-79702a1131e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.825047] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9660c8-abf8-4263-9d93-f866d3527766 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.825047] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd1df6f-c9a1-406e-871d-6df45e8fab92 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.825047] env[70013]: DEBUG nova.network.neutron [-] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.825047] env[70013]: DEBUG oslo_vmware.api [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230540, 'name': PowerOnVM_Task, 'duration_secs': 1.08556} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.825047] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 717.825047] env[70013]: DEBUG nova.compute.manager [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 717.840147] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be74946-8e5e-49b5-96e0-864b1351aff2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.840147] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d814091b-9bfc-4a89-8d63-1d663f71cee0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.840147] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.840147] env[70013]: DEBUG nova.compute.provider_tree [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.840147] env[70013]: DEBUG nova.network.neutron [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updated VIF entry in instance network info cache for port a0d11944-d177-42a5-ad2b-22c116396f8a. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 717.840727] env[70013]: DEBUG nova.network.neutron [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.841093] env[70013]: DEBUG nova.scheduler.client.report [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 717.858643] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230545, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.073896] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "refresh_cache-225c02d6-43df-4260-a428-14a6b1ed32cc" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.074236] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquired lock "refresh_cache-225c02d6-43df-4260-a428-14a6b1ed32cc" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.074469] env[70013]: DEBUG nova.network.neutron [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.127459] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230546, 'name': Rename_Task, 'duration_secs': 0.257688} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.127852] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 718.128927] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d258d93-3415-4341-a6b6-0e5b244e517f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.138243] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 718.138243] env[70013]: value = "task-4230548" [ 718.138243] env[70013]: _type = "Task" [ 718.138243] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.147816] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.218207] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230547, 'name': ReconfigVM_Task, 'duration_secs': 0.343486} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.218374] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Reconfigured VM instance instance-00000007 to attach disk [datastore1] 9a553d85-5086-46c5-8df4-451928e38ed9/9a553d85-5086-46c5-8df4-451928e38ed9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.218957] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28bf3f57-9d9e-4615-ab48-176f9527782c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.227508] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 718.227508] env[70013]: value = "task-4230549" [ 718.227508] env[70013]: _type = "Task" [ 718.227508] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.237390] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230549, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.306487] env[70013]: INFO nova.compute.manager [-] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Took 1.65 seconds to deallocate network for instance. [ 718.344460] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff03ce3b-95a5-4f30-bb0d-2d310fbb0559 req-0eb6a95c-4086-4505-8488-11ae16d6b6f2 service nova] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.346473] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.043s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.346473] env[70013]: DEBUG nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 718.352745] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.260s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.354090] env[70013]: INFO nova.compute.claims [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.366981] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230545, 'name': CreateVM_Task, 'duration_secs': 0.543718} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.367401] env[70013]: DEBUG nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 718.370142] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 718.374665] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.374866] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.375501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 718.376337] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b9edbfa-406b-4760-adb8-4536988454e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.381925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.390500] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 718.390500] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d51e22-2ebc-5f00-037e-9d1ac9a658c7" [ 718.390500] env[70013]: _type = "Task" [ 718.390500] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.401402] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d51e22-2ebc-5f00-037e-9d1ac9a658c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.407333] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 718.407790] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.407790] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.408013] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.408685] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.408685] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 718.408886] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 718.409161] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 718.409457] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 718.409735] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 718.410057] env[70013]: DEBUG nova.virt.hardware [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 718.411589] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d5c2e6-00b3-4b49-a960-4742026a1624 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.422350] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874dd529-d1fe-48c5-891b-954e117376f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.627909] env[70013]: DEBUG nova.network.neutron [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.650100] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230548, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.743870] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230549, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.818087] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.841714] env[70013]: DEBUG nova.network.neutron [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Updating instance_info_cache with network_info: [{"id": "aa55189e-9d4e-4fc1-9143-9b0ade9887b7", "address": "fa:16:3e:c4:f7:02", "network": {"id": "cb7bbbfa-16b1-447e-b356-6058fd9c50bf", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-263210435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94ebd7a191d4701a62223fbc2e1bdf8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa55189e-9d", "ovs_interfaceid": "aa55189e-9d4e-4fc1-9143-9b0ade9887b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.854510] env[70013]: DEBUG nova.compute.utils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 718.857497] env[70013]: DEBUG nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 718.857497] env[70013]: DEBUG nova.network.neutron [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.902078] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d51e22-2ebc-5f00-037e-9d1ac9a658c7, 'name': SearchDatastore_Task, 'duration_secs': 0.017235} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.905419] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.905419] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 718.905419] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.905419] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.905884] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 718.905884] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ce8754e-12df-4768-a1e8-9d45565feeb3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.915033] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 718.915331] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 718.916194] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-560cab7e-3600-4742-89e9-04f9c74d6a26 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.925371] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 718.925371] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cc9d9a-0fe6-2396-6f65-5e98f2f7a809" [ 718.925371] env[70013]: _type = "Task" [ 718.925371] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.938565] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cc9d9a-0fe6-2396-6f65-5e98f2f7a809, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.951543] env[70013]: DEBUG nova.policy [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd75bdd9e43e24b7caf23c8dd90f59791', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0583e7e4c4394e1f82da360b1c229b3e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 719.151018] env[70013]: DEBUG oslo_vmware.api [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230548, 'name': PowerOnVM_Task, 'duration_secs': 0.895085} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.151650] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 719.151854] env[70013]: INFO nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Took 8.48 seconds to spawn the instance on the hypervisor. [ 719.152060] env[70013]: DEBUG nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 719.153102] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05f06c2-3a6e-4b62-830b-2a27bcb684bd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.245827] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230549, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.349720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Releasing lock "refresh_cache-225c02d6-43df-4260-a428-14a6b1ed32cc" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.350617] env[70013]: DEBUG nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Instance network_info: |[{"id": "aa55189e-9d4e-4fc1-9143-9b0ade9887b7", "address": "fa:16:3e:c4:f7:02", "network": {"id": "cb7bbbfa-16b1-447e-b356-6058fd9c50bf", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-263210435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94ebd7a191d4701a62223fbc2e1bdf8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa55189e-9d", "ovs_interfaceid": "aa55189e-9d4e-4fc1-9143-9b0ade9887b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 719.351711] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:f7:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa55189e-9d4e-4fc1-9143-9b0ade9887b7', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 719.363044] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Creating folder: Project (e94ebd7a191d4701a62223fbc2e1bdf8). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 719.363658] env[70013]: DEBUG nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 719.366291] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d50a5d3e-7696-4796-9311-2462eb71db74 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.391069] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Created folder: Project (e94ebd7a191d4701a62223fbc2e1bdf8) in parent group-v836999. [ 719.391279] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Creating folder: Instances. Parent ref: group-v837028. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 719.391530] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f47c13e-4777-4350-a74c-5ba584c7a7f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.406287] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Created folder: Instances in parent group-v837028. [ 719.406609] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 719.406846] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 719.407106] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b765c627-3803-4e52-acc8-93f0de801a5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.428847] env[70013]: DEBUG nova.network.neutron [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Successfully created port: 9a809fa3-d1c4-4bff-b415-a9845cbdc5ad {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.438147] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 719.438147] env[70013]: value = "task-4230552" [ 719.438147] env[70013]: _type = "Task" [ 719.438147] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.447652] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cc9d9a-0fe6-2396-6f65-5e98f2f7a809, 'name': SearchDatastore_Task, 'duration_secs': 0.045491} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.452015] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5ebd560-afc1-4b9a-99fb-98419344ffd6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.458614] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230552, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.463299] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 719.463299] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52081210-f595-0d0d-1c7d-eb452cb0cdad" [ 719.463299] env[70013]: _type = "Task" [ 719.463299] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.480632] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52081210-f595-0d0d-1c7d-eb452cb0cdad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.553936] env[70013]: DEBUG nova.compute.manager [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Received event network-vif-plugged-aa55189e-9d4e-4fc1-9143-9b0ade9887b7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 719.554153] env[70013]: DEBUG oslo_concurrency.lockutils [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] Acquiring lock "225c02d6-43df-4260-a428-14a6b1ed32cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.554353] env[70013]: DEBUG oslo_concurrency.lockutils [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 719.554511] env[70013]: DEBUG oslo_concurrency.lockutils [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 719.554666] env[70013]: DEBUG nova.compute.manager [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] No waiting events found dispatching network-vif-plugged-aa55189e-9d4e-4fc1-9143-9b0ade9887b7 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 719.554821] env[70013]: WARNING nova.compute.manager [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Received unexpected event network-vif-plugged-aa55189e-9d4e-4fc1-9143-9b0ade9887b7 for instance with vm_state building and task_state spawning. [ 719.555134] env[70013]: DEBUG nova.compute.manager [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Received event network-changed-aa55189e-9d4e-4fc1-9143-9b0ade9887b7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 719.560700] env[70013]: DEBUG nova.compute.manager [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Refreshing instance network info cache due to event network-changed-aa55189e-9d4e-4fc1-9143-9b0ade9887b7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 719.560700] env[70013]: DEBUG oslo_concurrency.lockutils [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] Acquiring lock "refresh_cache-225c02d6-43df-4260-a428-14a6b1ed32cc" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.560826] env[70013]: DEBUG oslo_concurrency.lockutils [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] Acquired lock "refresh_cache-225c02d6-43df-4260-a428-14a6b1ed32cc" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.560934] env[70013]: DEBUG nova.network.neutron [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Refreshing network info cache for port aa55189e-9d4e-4fc1-9143-9b0ade9887b7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 719.682931] env[70013]: INFO nova.compute.manager [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Took 22.07 seconds to build instance. [ 719.748096] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230549, 'name': Rename_Task, 'duration_secs': 1.209194} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.748096] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 719.748289] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a30558b0-af3c-4a18-9466-1cfaf874ecfc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.760928] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 719.760928] env[70013]: value = "task-4230553" [ 719.760928] env[70013]: _type = "Task" [ 719.760928] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.771022] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.819590] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36fe4787-2810-4620-a606-e997b55b0a84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.830584] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b501d8-f638-444c-95ee-3f1f3f710919 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.875977] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e141237f-f62d-4e72-ad1d-28bd83dce6a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.888772] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f68893-0bc7-4922-b1ae-0ce5a090d763 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.905238] env[70013]: DEBUG nova.compute.provider_tree [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.915291] env[70013]: DEBUG nova.network.neutron [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Successfully updated port: a4ff3dbe-e842-4e4a-939e-e3f429ebad04 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 719.954671] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230552, 'name': CreateVM_Task, 'duration_secs': 0.391997} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.956165] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 719.956165] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.956165] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.956499] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 719.957099] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af7d3f60-b2d2-490d-b3c3-caa75d5dd4a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.964393] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 719.964393] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b02363-25d2-e306-58d6-986ae2ca6f6b" [ 719.964393] env[70013]: _type = "Task" [ 719.964393] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.991068] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b02363-25d2-e306-58d6-986ae2ca6f6b, 'name': SearchDatastore_Task, 'duration_secs': 0.01501} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.991068] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52081210-f595-0d0d-1c7d-eb452cb0cdad, 'name': SearchDatastore_Task, 'duration_secs': 0.014796} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.991068] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.991068] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.991356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.991356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.991356] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 28a56a75-9b30-4121-8252-a9e57287441c/28a56a75-9b30-4121-8252-a9e57287441c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 719.991356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.991506] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.991506] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b386d1ec-3361-4fbd-9365-12ea182c17a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.993748] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edd605f1-6021-438d-8372-3db3c6120d08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.012338] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 720.012338] env[70013]: value = "task-4230554" [ 720.012338] env[70013]: _type = "Task" [ 720.012338] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.022401] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 720.022645] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 720.024549] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e36f845d-2601-45bb-a73e-0e4c8180de8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.031486] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.036603] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 720.036603] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527ca0ae-919c-f63b-70f3-3315007e86d7" [ 720.036603] env[70013]: _type = "Task" [ 720.036603] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.050816] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527ca0ae-919c-f63b-70f3-3315007e86d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.086207] env[70013]: DEBUG nova.compute.manager [req-4f75ffad-c9fa-401b-adee-58c7419b8174 req-50cd2d73-4f25-45e3-a0ef-224a8d0f570e service nova] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Received event network-vif-deleted-7edceb22-219c-4108-abc8-dabed5a1e064 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 720.086908] env[70013]: DEBUG nova.compute.manager [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Stashing vm_state: active {{(pid=70013) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 720.190950] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e7821c81-1b2c-42a8-b291-c29e862ade36 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "cf8be667-da84-466b-9b7b-3bdfda6a8193" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.600s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.274240] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230553, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.322625] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "417f8c74-81cc-47d6-80ca-b61d03ebec50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.322969] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.323234] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "417f8c74-81cc-47d6-80ca-b61d03ebec50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.323506] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.323800] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.326457] env[70013]: INFO nova.compute.manager [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Terminating instance [ 720.384065] env[70013]: DEBUG nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 720.414371] env[70013]: DEBUG nova.scheduler.client.report [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 720.425432] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 720.425432] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.425432] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 720.425828] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.425828] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 720.425828] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 720.425828] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 720.426522] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 720.426734] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 720.426909] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 720.427161] env[70013]: DEBUG nova.virt.hardware [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 720.431498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.431498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.431498] env[70013]: DEBUG nova.network.neutron [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.436021] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9838f6-f016-412e-a18b-a4cc91a1cd7d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.452811] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbdd37b-9bcf-421b-b4ad-75b1559c7acd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.525776] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230554, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.550807] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527ca0ae-919c-f63b-70f3-3315007e86d7, 'name': SearchDatastore_Task, 'duration_secs': 0.012144} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.551906] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2744f8fd-2d3a-4b8f-bb5e-24fe15398ef6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.561599] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 720.561599] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5228ecc1-0afa-783a-14af-aae360b4141e" [ 720.561599] env[70013]: _type = "Task" [ 720.561599] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.577934] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5228ecc1-0afa-783a-14af-aae360b4141e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.614687] env[70013]: DEBUG nova.network.neutron [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Updated VIF entry in instance network info cache for port aa55189e-9d4e-4fc1-9143-9b0ade9887b7. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 720.614687] env[70013]: DEBUG nova.network.neutron [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Updating instance_info_cache with network_info: [{"id": "aa55189e-9d4e-4fc1-9143-9b0ade9887b7", "address": "fa:16:3e:c4:f7:02", "network": {"id": "cb7bbbfa-16b1-447e-b356-6058fd9c50bf", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-263210435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94ebd7a191d4701a62223fbc2e1bdf8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa55189e-9d", "ovs_interfaceid": "aa55189e-9d4e-4fc1-9143-9b0ade9887b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.618143] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.694806] env[70013]: DEBUG nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 720.777067] env[70013]: DEBUG oslo_vmware.api [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230553, 'name': PowerOnVM_Task, 'duration_secs': 0.55136} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.777346] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 720.779254] env[70013]: INFO nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Took 12.65 seconds to spawn the instance on the hypervisor. [ 720.779254] env[70013]: DEBUG nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 720.779254] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150e4a14-718d-4258-99d4-a7f519982fdd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.832164] env[70013]: DEBUG nova.compute.manager [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 720.834907] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.834907] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d9e926-80bd-4d4c-8d93-899b920d8d95 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.846388] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 720.847876] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb069947-58b0-45c5-a0bf-b512e5dc9368 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.857482] env[70013]: DEBUG oslo_vmware.api [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 720.857482] env[70013]: value = "task-4230555" [ 720.857482] env[70013]: _type = "Task" [ 720.857482] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.869484] env[70013]: DEBUG oslo_vmware.api [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.942926] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.943693] env[70013]: DEBUG nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 720.947702] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.617s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.974857] env[70013]: DEBUG nova.network.neutron [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.028707] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230554, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685432} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.029728] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 28a56a75-9b30-4121-8252-a9e57287441c/28a56a75-9b30-4121-8252-a9e57287441c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.032072] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.032072] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c3964b8a-97c2-4be1-9b05-ec39e9e68531 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.041151] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 721.041151] env[70013]: value = "task-4230556" [ 721.041151] env[70013]: _type = "Task" [ 721.041151] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.057458] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230556, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.079368] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5228ecc1-0afa-783a-14af-aae360b4141e, 'name': SearchDatastore_Task, 'duration_secs': 0.060965} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.079963] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.080312] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 225c02d6-43df-4260-a428-14a6b1ed32cc/225c02d6-43df-4260-a428-14a6b1ed32cc.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 721.080671] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e671660f-20a7-461e-abac-0c13c15c8b32 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.094791] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 721.094791] env[70013]: value = "task-4230557" [ 721.094791] env[70013]: _type = "Task" [ 721.094791] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.094791] env[70013]: DEBUG nova.network.neutron [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Successfully updated port: 9a809fa3-d1c4-4bff-b415-a9845cbdc5ad {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 721.106859] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230557, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.117256] env[70013]: DEBUG oslo_concurrency.lockutils [req-9613ea72-e0e9-40ac-a798-9674486b7e00 req-36222340-e76b-4c21-a47f-a36e57567e9f service nova] Releasing lock "refresh_cache-225c02d6-43df-4260-a428-14a6b1ed32cc" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.224263] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.305384] env[70013]: INFO nova.compute.manager [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Took 25.05 seconds to build instance. [ 721.366127] env[70013]: DEBUG nova.network.neutron [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Updating instance_info_cache with network_info: [{"id": "a4ff3dbe-e842-4e4a-939e-e3f429ebad04", "address": "fa:16:3e:27:95:6a", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ff3dbe-e8", "ovs_interfaceid": "a4ff3dbe-e842-4e4a-939e-e3f429ebad04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.378244] env[70013]: DEBUG oslo_vmware.api [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230555, 'name': PowerOffVM_Task, 'duration_secs': 0.423911} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.380679] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 721.380679] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 721.380679] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-240ff82f-9131-4a4b-9b92-0c7090216c69 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.469415] env[70013]: DEBUG nova.compute.utils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 721.474299] env[70013]: DEBUG nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 721.474299] env[70013]: DEBUG nova.network.neutron [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.475324] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 721.475452] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 721.475656] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Deleting the datastore file [datastore1] 417f8c74-81cc-47d6-80ca-b61d03ebec50 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 721.477068] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b93aec1-752f-4c51-8f70-3e7d78ff3965 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.490859] env[70013]: DEBUG oslo_vmware.api [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for the task: (returnval){ [ 721.490859] env[70013]: value = "task-4230559" [ 721.490859] env[70013]: _type = "Task" [ 721.490859] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.506926] env[70013]: DEBUG oslo_vmware.api [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230559, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.553603] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230556, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097823} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.554080] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.554984] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d547d0-e55b-4d31-9794-9952c4c60f0f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.578095] env[70013]: INFO nova.compute.manager [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Rebuilding instance [ 721.597212] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] 28a56a75-9b30-4121-8252-a9e57287441c/28a56a75-9b30-4121-8252-a9e57287441c.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.600403] env[70013]: DEBUG nova.policy [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac21e3c87f0d4c6fbff58c8bd524b181', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '719bc6e70189480dbef8b541a7a63d36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 721.605435] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "refresh_cache-64308d7c-b63b-4fa6-b235-2f0586a888cf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.605435] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquired lock "refresh_cache-64308d7c-b63b-4fa6-b235-2f0586a888cf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.605719] env[70013]: DEBUG nova.network.neutron [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.608077] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8af05e07-4bb6-4a5d-918e-178097090c1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.650051] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230557, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.654981] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 721.654981] env[70013]: value = "task-4230560" [ 721.654981] env[70013]: _type = "Task" [ 721.654981] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.673069] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230560, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.690664] env[70013]: DEBUG nova.compute.manager [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 721.692346] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00771171-7179-4cf6-99dd-21747464e557 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.736127] env[70013]: DEBUG nova.network.neutron [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.810351] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b8a3ac39-f38e-49dd-8060-ec950b00747e tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "9a553d85-5086-46c5-8df4-451928e38ed9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.604s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.875496] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.875496] env[70013]: DEBUG nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance network_info: |[{"id": "a4ff3dbe-e842-4e4a-939e-e3f429ebad04", "address": "fa:16:3e:27:95:6a", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ff3dbe-e8", "ovs_interfaceid": "a4ff3dbe-e842-4e4a-939e-e3f429ebad04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 721.875652] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:95:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a4ff3dbe-e842-4e4a-939e-e3f429ebad04', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 721.885352] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating folder: Project (75ec7d6e9ea54c06bd3b7605f16b391a). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 721.889608] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9538c544-e8f3-4d8a-ab6b-ff20ad57ec16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.905469] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Created folder: Project (75ec7d6e9ea54c06bd3b7605f16b391a) in parent group-v836999. [ 721.906106] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating folder: Instances. Parent ref: group-v837031. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 721.906220] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5eb235c0-7b24-43dc-bc82-7ece946d253a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.918405] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Created folder: Instances in parent group-v837031. [ 721.919143] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 721.919314] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 721.919542] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c02f6bb9-a67c-4a14-9479-98514326132c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.945617] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 721.945617] env[70013]: value = "task-4230563" [ 721.945617] env[70013]: _type = "Task" [ 721.945617] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.957380] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230563, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.974918] env[70013]: INFO nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating resource usage from migration 2b86447b-3ba9-457a-94d4-7269f0b254ab [ 721.977036] env[70013]: DEBUG nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 721.998602] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2b91ec52-2d32-4f86-acf2-f74f661f65b5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 721.999489] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 98a0199a-4209-4faf-adf5-7ae33b099d20 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004153] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance e8998cab-533b-46b1-a1ea-5eb5db7d6d99 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 722.004153] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9c5f2422-820e-454b-98be-695e3ac6d1e3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 722.004153] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 417f8c74-81cc-47d6-80ca-b61d03ebec50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004153] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a553d85-5086-46c5-8df4-451928e38ed9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004300] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance cf8be667-da84-466b-9b7b-3bdfda6a8193 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004300] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 28a56a75-9b30-4121-8252-a9e57287441c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004300] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 225c02d6-43df-4260-a428-14a6b1ed32cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004300] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 63c3b57c-022a-4eee-b215-6dd89142e659 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004414] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 64308d7c-b63b-4fa6-b235-2f0586a888cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.004414] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 722.008895] env[70013]: DEBUG oslo_vmware.api [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Task: {'id': task-4230559, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.414518} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.009744] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 722.010658] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 722.010878] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.011087] env[70013]: INFO nova.compute.manager [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Took 1.18 seconds to destroy the instance on the hypervisor. [ 722.011448] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 722.011549] env[70013]: DEBUG nova.compute.manager [-] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 722.011632] env[70013]: DEBUG nova.network.neutron [-] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.014682] env[70013]: DEBUG nova.network.neutron [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Updating instance_info_cache with network_info: [{"id": "9a809fa3-d1c4-4bff-b415-a9845cbdc5ad", "address": "fa:16:3e:aa:15:72", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a809fa3-d1", "ovs_interfaceid": "9a809fa3-d1c4-4bff-b415-a9845cbdc5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.118682] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230557, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67258} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.119203] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 225c02d6-43df-4260-a428-14a6b1ed32cc/225c02d6-43df-4260-a428-14a6b1ed32cc.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 722.119811] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 722.120737] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca3ab6ca-12c1-4eee-8eac-e4f6886c8bce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.129605] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 722.129605] env[70013]: value = "task-4230564" [ 722.129605] env[70013]: _type = "Task" [ 722.129605] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.144500] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230564, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.169070] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230560, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.313707] env[70013]: DEBUG nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 722.353236] env[70013]: DEBUG nova.network.neutron [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Successfully created port: 0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.459243] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230563, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.513484] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance fc0b7f87-8a36-440e-b158-30b9d9920572 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 722.521020] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Releasing lock "refresh_cache-64308d7c-b63b-4fa6-b235-2f0586a888cf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.521020] env[70013]: DEBUG nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Instance network_info: |[{"id": "9a809fa3-d1c4-4bff-b415-a9845cbdc5ad", "address": "fa:16:3e:aa:15:72", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a809fa3-d1", "ovs_interfaceid": "9a809fa3-d1c4-4bff-b415-a9845cbdc5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 722.521342] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:15:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a809fa3-d1c4-4bff-b415-a9845cbdc5ad', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.533854] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Creating folder: Project (0583e7e4c4394e1f82da360b1c229b3e). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.535048] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ab49d2e-ef8e-48d5-b92b-ad9bac5f2777 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.546391] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Created folder: Project (0583e7e4c4394e1f82da360b1c229b3e) in parent group-v836999. [ 722.546671] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Creating folder: Instances. Parent ref: group-v837034. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.546828] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1356c2f1-5426-4938-82b5-ebdde25c50a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.564086] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Created folder: Instances in parent group-v837034. [ 722.564605] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 722.564873] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 722.566094] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a936e8d4-069b-4f79-9b35-ff031f7a9e72 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.589119] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.589119] env[70013]: value = "task-4230567" [ 722.589119] env[70013]: _type = "Task" [ 722.589119] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.598046] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230567, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.641961] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230564, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094913} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.642823] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.645472] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae26107-60ab-454a-9ffb-3a5940a382fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.671451] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 225c02d6-43df-4260-a428-14a6b1ed32cc/225c02d6-43df-4260-a428-14a6b1ed32cc.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.674270] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd0a1f3e-c91e-4ea5-add4-73a426ae64b0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.695386] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230560, 'name': ReconfigVM_Task, 'duration_secs': 0.74446} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.697053] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Reconfigured VM instance instance-00000009 to attach disk [datastore2] 28a56a75-9b30-4121-8252-a9e57287441c/28a56a75-9b30-4121-8252-a9e57287441c.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.697923] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 722.697923] env[70013]: value = "task-4230568" [ 722.697923] env[70013]: _type = "Task" [ 722.697923] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.698507] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-371a619c-3967-4022-8348-36de50caf724 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.711084] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230568, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.712507] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 722.712507] env[70013]: value = "task-4230569" [ 722.712507] env[70013]: _type = "Task" [ 722.712507] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.713379] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 722.713598] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f508b74-faf6-4a01-963b-bbdc76d7e4ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.730555] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230569, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.730956] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 722.730956] env[70013]: value = "task-4230570" [ 722.730956] env[70013]: _type = "Task" [ 722.730956] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.737508] env[70013]: DEBUG nova.compute.manager [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Received event network-vif-plugged-9a809fa3-d1c4-4bff-b415-a9845cbdc5ad {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 722.737626] env[70013]: DEBUG oslo_concurrency.lockutils [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] Acquiring lock "64308d7c-b63b-4fa6-b235-2f0586a888cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.737799] env[70013]: DEBUG oslo_concurrency.lockutils [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 722.738926] env[70013]: DEBUG oslo_concurrency.lockutils [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 722.738926] env[70013]: DEBUG nova.compute.manager [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] No waiting events found dispatching network-vif-plugged-9a809fa3-d1c4-4bff-b415-a9845cbdc5ad {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 722.738926] env[70013]: WARNING nova.compute.manager [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Received unexpected event network-vif-plugged-9a809fa3-d1c4-4bff-b415-a9845cbdc5ad for instance with vm_state building and task_state spawning. [ 722.738926] env[70013]: DEBUG nova.compute.manager [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Received event network-changed-9a809fa3-d1c4-4bff-b415-a9845cbdc5ad {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 722.738926] env[70013]: DEBUG nova.compute.manager [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Refreshing instance network info cache due to event network-changed-9a809fa3-d1c4-4bff-b415-a9845cbdc5ad. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 722.739415] env[70013]: DEBUG oslo_concurrency.lockutils [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] Acquiring lock "refresh_cache-64308d7c-b63b-4fa6-b235-2f0586a888cf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.739415] env[70013]: DEBUG oslo_concurrency.lockutils [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] Acquired lock "refresh_cache-64308d7c-b63b-4fa6-b235-2f0586a888cf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.739415] env[70013]: DEBUG nova.network.neutron [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Refreshing network info cache for port 9a809fa3-d1c4-4bff-b415-a9845cbdc5ad {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.746961] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.858955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.962647] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230563, 'name': CreateVM_Task, 'duration_secs': 0.557465} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.962984] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 722.963827] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.964104] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.964833] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 722.964833] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec6156c9-bd62-4fde-85e1-4f49f96235bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.972338] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 722.972338] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5210284f-08ce-323f-aeae-862224f3ab32" [ 722.972338] env[70013]: _type = "Task" [ 722.972338] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.982096] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5210284f-08ce-323f-aeae-862224f3ab32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.989723] env[70013]: DEBUG nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 723.018904] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance fc473d40-b57b-437e-9511-58a0ba700a69 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 723.029045] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 723.029281] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.030154] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 723.030154] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.030154] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 723.030154] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 723.032265] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 723.032619] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 723.032685] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 723.032914] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 723.033516] env[70013]: DEBUG nova.virt.hardware [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 723.033938] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456e6d4d-8e34-4f2d-ba7e-c61ae0bc15c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.049031] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c2958e-189d-49f3-8fa6-4fc5ad1bfea8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.104331] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230567, 'name': CreateVM_Task, 'duration_secs': 0.487454} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.104922] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 723.105672] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.216191] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230568, 'name': ReconfigVM_Task, 'duration_secs': 0.511514} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.226045] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 225c02d6-43df-4260-a428-14a6b1ed32cc/225c02d6-43df-4260-a428-14a6b1ed32cc.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.230730] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0689ccfa-ec29-431c-a762-48a09a25f011 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.250250] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230569, 'name': Rename_Task, 'duration_secs': 0.262764} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.252980] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.253610] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 723.253610] env[70013]: value = "task-4230571" [ 723.253610] env[70013]: _type = "Task" [ 723.253610] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.254114] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24ff1758-908c-47fd-88c8-beedcecc7fe0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.266200] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230570, 'name': PowerOffVM_Task, 'duration_secs': 0.173054} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.266200] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 723.266200] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 723.267585] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ee4bc5-d80d-4a5f-8ae0-d7dab631eaa7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.276829] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230571, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.277221] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 723.277221] env[70013]: value = "task-4230572" [ 723.277221] env[70013]: _type = "Task" [ 723.277221] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.283755] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 723.284484] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-348cd895-9562-4968-ba1c-a972fe83aa67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.291686] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.317992] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 723.318336] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 723.318448] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Deleting the datastore file [datastore1] 98a0199a-4209-4faf-adf5-7ae33b099d20 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 723.318779] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe3aad11-dae6-4ae8-a46b-9f05207b08f9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.327872] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 723.327872] env[70013]: value = "task-4230574" [ 723.327872] env[70013]: _type = "Task" [ 723.327872] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.339363] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230574, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.433198] env[70013]: DEBUG nova.compute.manager [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Received event network-vif-plugged-a4ff3dbe-e842-4e4a-939e-e3f429ebad04 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 723.433483] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] Acquiring lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.433841] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] Lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.434403] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] Lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.437066] env[70013]: DEBUG nova.compute.manager [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] No waiting events found dispatching network-vif-plugged-a4ff3dbe-e842-4e4a-939e-e3f429ebad04 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 723.437066] env[70013]: WARNING nova.compute.manager [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Received unexpected event network-vif-plugged-a4ff3dbe-e842-4e4a-939e-e3f429ebad04 for instance with vm_state building and task_state spawning. [ 723.437066] env[70013]: DEBUG nova.compute.manager [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Received event network-changed-a4ff3dbe-e842-4e4a-939e-e3f429ebad04 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 723.437066] env[70013]: DEBUG nova.compute.manager [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Refreshing instance network info cache due to event network-changed-a4ff3dbe-e842-4e4a-939e-e3f429ebad04. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 723.437066] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] Acquiring lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.437296] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] Acquired lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.437359] env[70013]: DEBUG nova.network.neutron [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Refreshing network info cache for port a4ff3dbe-e842-4e4a-939e-e3f429ebad04 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 723.492314] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5210284f-08ce-323f-aeae-862224f3ab32, 'name': SearchDatastore_Task, 'duration_secs': 0.013858} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.492495] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.492712] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.492967] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.493146] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.493331] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.493922] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.494087] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 723.494384] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e49489ff-4462-4741-b3da-3e94fcf95194 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.496472] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99a1dd8d-4b58-4750-b554-12bd40d0e8e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.505122] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 723.505122] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ed73c0-2f09-6d10-3811-328e42b6d051" [ 723.505122] env[70013]: _type = "Task" [ 723.505122] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.512353] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.512539] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 723.513808] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b8cd179-cc4e-4ae6-872f-b618c1cae303 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.520995] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ed73c0-2f09-6d10-3811-328e42b6d051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.522202] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 723.525803] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 723.525803] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bc8e55-7411-6bc2-0470-b311aa06251b" [ 723.525803] env[70013]: _type = "Task" [ 723.525803] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.536774] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bc8e55-7411-6bc2-0470-b311aa06251b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.556037] env[70013]: DEBUG nova.network.neutron [-] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.691032] env[70013]: DEBUG nova.network.neutron [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Updated VIF entry in instance network info cache for port 9a809fa3-d1c4-4bff-b415-a9845cbdc5ad. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 723.691032] env[70013]: DEBUG nova.network.neutron [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Updating instance_info_cache with network_info: [{"id": "9a809fa3-d1c4-4bff-b415-a9845cbdc5ad", "address": "fa:16:3e:aa:15:72", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a809fa3-d1", "ovs_interfaceid": "9a809fa3-d1c4-4bff-b415-a9845cbdc5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.738039] env[70013]: INFO nova.compute.manager [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Rebuilding instance [ 723.769452] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230571, 'name': Rename_Task, 'duration_secs': 0.239767} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.772474] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.772956] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cdc60c4b-4fd8-4ec5-924b-0326d7fd2e0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.792935] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 723.792935] env[70013]: value = "task-4230575" [ 723.792935] env[70013]: _type = "Task" [ 723.792935] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.804816] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230572, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.805284] env[70013]: DEBUG nova.compute.manager [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 723.806614] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a0e31e-b0fb-4be1-8f62-5a5f40429f2f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.814053] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230575, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.842127] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230574, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155761} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.842127] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 723.842127] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 723.842127] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.019827] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ed73c0-2f09-6d10-3811-328e42b6d051, 'name': SearchDatastore_Task, 'duration_secs': 0.021352} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.020165] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.020444] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 724.021043] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.026764] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 724.054535] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bc8e55-7411-6bc2-0470-b311aa06251b, 'name': SearchDatastore_Task, 'duration_secs': 0.017425} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.055855] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e13b4862-7188-4b96-b42c-a4bb1898a7f4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.059404] env[70013]: INFO nova.compute.manager [-] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Took 2.05 seconds to deallocate network for instance. [ 724.066741] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 724.066741] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52980b05-0b87-70b3-ba80-c33964e46f21" [ 724.066741] env[70013]: _type = "Task" [ 724.066741] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.078811] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52980b05-0b87-70b3-ba80-c33964e46f21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.195127] env[70013]: DEBUG oslo_concurrency.lockutils [req-46b89dab-b85e-4e3f-a3c9-f7ba663e4031 req-b566571b-68a5-4678-b694-2b8a4bf14980 service nova] Releasing lock "refresh_cache-64308d7c-b63b-4fa6-b235-2f0586a888cf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.295997] env[70013]: DEBUG oslo_vmware.api [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230572, 'name': PowerOnVM_Task, 'duration_secs': 0.872734} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.299765] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.300186] env[70013]: INFO nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Took 11.03 seconds to spawn the instance on the hypervisor. [ 724.300473] env[70013]: DEBUG nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 724.305091] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24258cad-89ba-461c-b8e0-34e28ec25387 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.317677] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230575, 'name': PowerOnVM_Task} progress is 87%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.432722] env[70013]: DEBUG nova.network.neutron [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Updated VIF entry in instance network info cache for port a4ff3dbe-e842-4e4a-939e-e3f429ebad04. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 724.433179] env[70013]: DEBUG nova.network.neutron [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Updating instance_info_cache with network_info: [{"id": "a4ff3dbe-e842-4e4a-939e-e3f429ebad04", "address": "fa:16:3e:27:95:6a", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ff3dbe-e8", "ovs_interfaceid": "a4ff3dbe-e842-4e4a-939e-e3f429ebad04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.539655] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance b23b4e67-67a6-4117-9a62-3c0dcd25e828 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 724.539655] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Migration 2b86447b-3ba9-457a-94d4-7269f0b254ab is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1742}} [ 724.539655] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 3b29c99a-0d55-40b1-a155-199b1ced2146 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 724.539954] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 724.539954] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 724.569727] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 724.589552] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52980b05-0b87-70b3-ba80-c33964e46f21, 'name': SearchDatastore_Task, 'duration_secs': 0.017476} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.591836] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.591836] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 63c3b57c-022a-4eee-b215-6dd89142e659/63c3b57c-022a-4eee-b215-6dd89142e659.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 724.591836] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.591836] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 724.592051] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bbf8b7b-2735-4f8a-af6a-67e9cafc2b67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.597122] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8ca3c42-5c61-4bec-8134-5550bce8886b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.610536] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 724.610536] env[70013]: value = "task-4230576" [ 724.610536] env[70013]: _type = "Task" [ 724.610536] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.612531] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 724.613098] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 724.618698] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97cc913b-4eaa-4b37-a7d7-e61da13584fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.633777] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 724.633777] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a5e396-b39c-bed8-0230-9c9dd468e752" [ 724.633777] env[70013]: _type = "Task" [ 724.633777] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.634519] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.644793] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a5e396-b39c-bed8-0230-9c9dd468e752, 'name': SearchDatastore_Task, 'duration_secs': 0.011284} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.649363] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eab9133c-0564-49a5-a3c7-0d2fcc6888fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.656816] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 724.656816] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5252749b-de4a-a79d-6626-e2b5652bdd2e" [ 724.656816] env[70013]: _type = "Task" [ 724.656816] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.672847] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5252749b-de4a-a79d-6626-e2b5652bdd2e, 'name': SearchDatastore_Task, 'duration_secs': 0.010234} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.676846] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.677087] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 64308d7c-b63b-4fa6-b235-2f0586a888cf/64308d7c-b63b-4fa6-b235-2f0586a888cf.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 724.677802] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e31337f-11cc-47a7-80f6-8fb903e4caa5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.687349] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 724.687349] env[70013]: value = "task-4230577" [ 724.687349] env[70013]: _type = "Task" [ 724.687349] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.703798] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.814564] env[70013]: DEBUG oslo_vmware.api [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230575, 'name': PowerOnVM_Task, 'duration_secs': 0.847221} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.814874] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.815207] env[70013]: INFO nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Took 8.95 seconds to spawn the instance on the hypervisor. [ 724.815301] env[70013]: DEBUG nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 724.816478] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cf4258-9969-4396-a90d-715071513420 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.837456] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 724.839880] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01739135-cf54-47c4-a530-8e5ab23cc468 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.842118] env[70013]: INFO nova.compute.manager [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Took 27.19 seconds to build instance. [ 724.860592] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 724.860592] env[70013]: value = "task-4230578" [ 724.860592] env[70013]: _type = "Task" [ 724.860592] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.884730] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.905070] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 724.905481] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.905689] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 724.905967] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.906377] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 724.906568] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 724.906800] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 724.906965] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 724.907171] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 724.908578] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 724.908860] env[70013]: DEBUG nova.virt.hardware [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 724.911303] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85708dc-605d-4d90-9a85-71a439d57ac2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.926253] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5861530b-ee71-4690-a9de-11806adf01f2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.934043] env[70013]: DEBUG nova.network.neutron [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Successfully updated port: 0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 724.935509] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a183eb1-aa0c-4ef3-9eee-f9728fa8ef09 req-8daa649e-fbea-48f9-b89e-5f2c8be53214 service nova] Releasing lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.950396] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.956370] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 724.960138] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 724.960688] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ec33db3-52a0-4c30-8212-e9239fc56679 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.978169] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1359219-f269-4fcb-9c34-1f79e5f9620c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.983418] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.983418] env[70013]: value = "task-4230579" [ 724.983418] env[70013]: _type = "Task" [ 724.983418] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.993877] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1de328e-7f44-4cf0-84ff-8f8f586a2b99 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.000240] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230579, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.035760] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556ff3fa-60bd-4982-86ce-8c8de08ae12b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.047237] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302d3371-561a-40f6-a751-f6911031bb40 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.071359] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.091476] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "9a553d85-5086-46c5-8df4-451928e38ed9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.091988] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "9a553d85-5086-46c5-8df4-451928e38ed9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.092243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "9a553d85-5086-46c5-8df4-451928e38ed9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.092516] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "9a553d85-5086-46c5-8df4-451928e38ed9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.093338] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "9a553d85-5086-46c5-8df4-451928e38ed9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.095622] env[70013]: INFO nova.compute.manager [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Terminating instance [ 725.127176] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230576, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.207028] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230577, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.343901] env[70013]: INFO nova.compute.manager [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Took 25.91 seconds to build instance. [ 725.347854] env[70013]: DEBUG oslo_concurrency.lockutils [None req-852c9458-355b-49b9-ad95-9b07529624b4 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.704s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.376792] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230578, 'name': PowerOffVM_Task, 'duration_secs': 0.446629} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.376792] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 725.376792] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.377539] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b843a074-e9db-4925-9e50-28ad2689d914 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.388072] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 725.388365] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-784bb9db-3b84-46fe-a287-bf9b99413ccf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.423857] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 725.424102] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 725.424298] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Deleting the datastore file [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 725.424581] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dbcbc16a-bd59-4994-bf40-087cb4966ca7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.434374] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 725.434374] env[70013]: value = "task-4230581" [ 725.434374] env[70013]: _type = "Task" [ 725.434374] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.435062] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.435254] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.435432] env[70013]: DEBUG nova.network.neutron [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.450480] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230581, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.499776] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230579, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.580245] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 725.603759] env[70013]: DEBUG nova.compute.manager [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 725.603759] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.603759] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4faa7a41-e77f-403d-aeaa-ddf9422f22f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.617721] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 725.621955] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94a0a32f-9711-4a45-860a-7abf304d93a3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.631582] env[70013]: DEBUG oslo_vmware.api [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 725.631582] env[70013]: value = "task-4230582" [ 725.631582] env[70013]: _type = "Task" [ 725.631582] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.634731] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230576, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536214} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.638832] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 63c3b57c-022a-4eee-b215-6dd89142e659/63c3b57c-022a-4eee-b215-6dd89142e659.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 725.639536] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 725.639536] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71d8a445-7f3f-4bb1-b39b-c665fba8f6e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.647655] env[70013]: DEBUG oslo_vmware.api [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230582, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.649302] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 725.649302] env[70013]: value = "task-4230583" [ 725.649302] env[70013]: _type = "Task" [ 725.649302] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.663024] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230583, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.699059] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230577, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.866337} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.700182] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 64308d7c-b63b-4fa6-b235-2f0586a888cf/64308d7c-b63b-4fa6-b235-2f0586a888cf.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 725.700182] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 725.700182] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02b8826c-a084-4ed4-a857-6262e2c44f79 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.707838] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 725.707838] env[70013]: value = "task-4230584" [ 725.707838] env[70013]: _type = "Task" [ 725.707838] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.719219] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230584, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.851245] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c11520b-c679-40f7-baa7-c68d13316c0d tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.430s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.950376] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230581, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182192} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.951103] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 725.954406] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 725.954406] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.999690] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230579, 'name': CreateVM_Task, 'duration_secs': 0.673391} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.999877] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 726.003227] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.003227] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.003227] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 726.003227] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-add902dd-62db-4487-a529-d0a16dbb9c2e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.008019] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 726.008019] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52398d5c-bde3-504b-92aa-c842abb12a94" [ 726.008019] env[70013]: _type = "Task" [ 726.008019] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.017634] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52398d5c-bde3-504b-92aa-c842abb12a94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.039234] env[70013]: DEBUG nova.network.neutron [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.088026] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 726.088026] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.138s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.088026] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.504s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.088026] env[70013]: INFO nova.compute.claims [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.090870] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 726.091545] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Getting list of instances from cluster (obj){ [ 726.091545] env[70013]: value = "domain-c8" [ 726.091545] env[70013]: _type = "ClusterComputeResource" [ 726.091545] env[70013]: } {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 726.093159] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837546f5-837f-4059-8bc7-dea1eade660e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.118235] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Got total of 7 instances {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 726.131018] env[70013]: DEBUG nova.compute.manager [req-a5e28419-ef57-4885-aa57-47e8d91bed64 req-dd492d2a-5368-4498-8006-ca4979ccc0b7 service nova] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Received event network-vif-deleted-21a0b03c-bc07-44c7-b9e4-503338c106cd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 726.144468] env[70013]: DEBUG oslo_vmware.api [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230582, 'name': PowerOffVM_Task, 'duration_secs': 0.470147} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.144987] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 726.145206] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 726.145485] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b45cd1d9-422a-4a85-afce-6e7cfcedd2fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.159812] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230583, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094724} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.161184] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 726.161184] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c453cd-27a4-402f-ba20-a6ac45673f4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.192764] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 63c3b57c-022a-4eee-b215-6dd89142e659/63c3b57c-022a-4eee-b215-6dd89142e659.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 726.195959] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d0d2b43-836d-4edc-9636-3b55391b1525 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.221784] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230584, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102632} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.223874] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 726.224278] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 726.224278] env[70013]: value = "task-4230586" [ 726.224278] env[70013]: _type = "Task" [ 726.224278] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.225045] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e244cb9-3320-48ce-8849-af33004b7bb3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.237850] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230586, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.259737] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Reconfiguring VM instance instance-0000000c to attach disk [datastore2] 64308d7c-b63b-4fa6-b235-2f0586a888cf/64308d7c-b63b-4fa6-b235-2f0586a888cf.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 726.259900] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5d221c2-5ebd-4adb-a9d9-3d5578f30bbe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.275782] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 726.276020] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 726.276190] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Deleting the datastore file [datastore1] 9a553d85-5086-46c5-8df4-451928e38ed9 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 726.277173] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66d0e6dd-8c2a-42f3-9094-1ab85b0f5681 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.282067] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 726.282067] env[70013]: value = "task-4230587" [ 726.282067] env[70013]: _type = "Task" [ 726.282067] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.286803] env[70013]: DEBUG oslo_vmware.api [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for the task: (returnval){ [ 726.286803] env[70013]: value = "task-4230588" [ 726.286803] env[70013]: _type = "Task" [ 726.286803] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.293650] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230587, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.299036] env[70013]: DEBUG oslo_vmware.api [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230588, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.519680] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52398d5c-bde3-504b-92aa-c842abb12a94, 'name': SearchDatastore_Task, 'duration_secs': 0.010372} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.520213] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.520213] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.520493] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.520646] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.521304] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.521304] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9b79f3c-6e18-4ea1-98a0-7074c098dd01 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.530743] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.530946] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 726.531740] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d410bce2-087f-48e9-af7e-127b5a47c122 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.537879] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 726.537879] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52151713-df13-e069-e672-820cf43622c7" [ 726.537879] env[70013]: _type = "Task" [ 726.537879] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.546820] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52151713-df13-e069-e672-820cf43622c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.567510] env[70013]: DEBUG nova.network.neutron [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updating instance_info_cache with network_info: [{"id": "0da6d475-040d-43e8-9139-88652285a5fe", "address": "fa:16:3e:1e:a7:37", "network": {"id": "1c5012af-830c-4001-9013-9e8cf32698bb", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-222945680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "719bc6e70189480dbef8b541a7a63d36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0da6d475-04", "ovs_interfaceid": "0da6d475-040d-43e8-9139-88652285a5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.741494] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230586, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.757502] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "3e938bb3-54a6-49f8-809d-d5aee0349eae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.757502] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.797712] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230587, 'name': ReconfigVM_Task, 'duration_secs': 0.443424} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.798656] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Reconfigured VM instance instance-0000000c to attach disk [datastore2] 64308d7c-b63b-4fa6-b235-2f0586a888cf/64308d7c-b63b-4fa6-b235-2f0586a888cf.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 726.799655] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d747427-cf2f-4122-afdd-da0af110bc4f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.805602] env[70013]: DEBUG oslo_vmware.api [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Task: {'id': task-4230588, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229423} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.806441] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 726.806562] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 726.806773] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.807022] env[70013]: INFO nova.compute.manager [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Took 1.20 seconds to destroy the instance on the hypervisor. [ 726.807409] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 726.807715] env[70013]: DEBUG nova.compute.manager [-] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 726.807854] env[70013]: DEBUG nova.network.neutron [-] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.812017] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 726.812017] env[70013]: value = "task-4230589" [ 726.812017] env[70013]: _type = "Task" [ 726.812017] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.828910] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230589, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.993125] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 726.993125] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.993125] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 726.993426] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.993698] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 726.993816] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 726.994701] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 726.994903] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 726.995154] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 726.995526] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 726.995732] env[70013]: DEBUG nova.virt.hardware [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 726.997391] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0090dc6b-5023-4c6e-8db5-85454bae2b70 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.006753] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7002f853-773e-4ecf-b2a8-3de4ce66900e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.025373] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.034249] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 727.034895] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 727.038885] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3049e5e9-592a-48d8-8f0a-f1f1d88f1117 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.065965] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52151713-df13-e069-e672-820cf43622c7, 'name': SearchDatastore_Task, 'duration_secs': 0.018094} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.069214] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.069214] env[70013]: value = "task-4230590" [ 727.069214] env[70013]: _type = "Task" [ 727.069214] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.069214] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-906f0ee9-0319-43cd-815b-77d40059516f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.075384] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Releasing lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.075790] env[70013]: DEBUG nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Instance network_info: |[{"id": "0da6d475-040d-43e8-9139-88652285a5fe", "address": "fa:16:3e:1e:a7:37", "network": {"id": "1c5012af-830c-4001-9013-9e8cf32698bb", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-222945680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "719bc6e70189480dbef8b541a7a63d36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0da6d475-04", "ovs_interfaceid": "0da6d475-040d-43e8-9139-88652285a5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 727.080932] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:a7:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24b91e3b-50f1-4a16-b929-942e6b31b2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0da6d475-040d-43e8-9139-88652285a5fe', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.089987] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Creating folder: Project (719bc6e70189480dbef8b541a7a63d36). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.091614] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe32144c-6875-4efe-98ab-9d147b1c02f5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.095705] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 727.095705] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522cdb9d-d789-ebe6-87de-75f3ab77d2a6" [ 727.095705] env[70013]: _type = "Task" [ 727.095705] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.109509] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230590, 'name': CreateVM_Task} progress is 15%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.111933] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Created folder: Project (719bc6e70189480dbef8b541a7a63d36) in parent group-v836999. [ 727.112178] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Creating folder: Instances. Parent ref: group-v837039. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.112867] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-610f35b0-ea52-44bf-8d9d-4c28564591ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.119353] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522cdb9d-d789-ebe6-87de-75f3ab77d2a6, 'name': SearchDatastore_Task, 'duration_secs': 0.0104} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.119642] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.119913] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 727.120239] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52769d71-76b6-46bd-b462-54b63681c2ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.124213] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Created folder: Instances in parent group-v837039. [ 727.124505] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 727.125217] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 727.125478] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59cda1eb-3282-4a0b-9cb2-aa36e9529dad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.147933] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 727.147933] env[70013]: value = "task-4230593" [ 727.147933] env[70013]: _type = "Task" [ 727.147933] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.153535] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.153535] env[70013]: value = "task-4230594" [ 727.153535] env[70013]: _type = "Task" [ 727.153535] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.161586] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230593, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.173614] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230594, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.184253] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea689c5e-7942-4663-a907-0d8e2be01ef7 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "interface-225c02d6-43df-4260-a428-14a6b1ed32cc-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.184253] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea689c5e-7942-4663-a907-0d8e2be01ef7 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "interface-225c02d6-43df-4260-a428-14a6b1ed32cc-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.184377] env[70013]: DEBUG nova.objects.instance [None req-ea689c5e-7942-4663-a907-0d8e2be01ef7 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lazy-loading 'flavor' on Instance uuid 225c02d6-43df-4260-a428-14a6b1ed32cc {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.250644] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230586, 'name': ReconfigVM_Task, 'duration_secs': 0.613007} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.251440] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 63c3b57c-022a-4eee-b215-6dd89142e659/63c3b57c-022a-4eee-b215-6dd89142e659.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 727.251708] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7a490ce-fc64-47a0-b5ac-32aa654a9e73 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.261344] env[70013]: DEBUG nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 727.265970] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 727.265970] env[70013]: value = "task-4230595" [ 727.265970] env[70013]: _type = "Task" [ 727.265970] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.285522] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230595, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.334709] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230589, 'name': Rename_Task, 'duration_secs': 0.150404} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.335196] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 727.335945] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba3e367e-c5be-4830-934f-318e3495e3cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.351103] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 727.351103] env[70013]: value = "task-4230596" [ 727.351103] env[70013]: _type = "Task" [ 727.351103] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.371598] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.378707] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.380200] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.497369] env[70013]: DEBUG nova.compute.manager [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Received event network-vif-plugged-0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 727.497766] env[70013]: DEBUG oslo_concurrency.lockutils [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] Acquiring lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.498276] env[70013]: DEBUG oslo_concurrency.lockutils [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.498276] env[70013]: DEBUG oslo_concurrency.lockutils [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.499843] env[70013]: DEBUG nova.compute.manager [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] No waiting events found dispatching network-vif-plugged-0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 727.499843] env[70013]: WARNING nova.compute.manager [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Received unexpected event network-vif-plugged-0da6d475-040d-43e8-9139-88652285a5fe for instance with vm_state building and task_state spawning. [ 727.499843] env[70013]: DEBUG nova.compute.manager [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Received event network-changed-0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 727.500054] env[70013]: DEBUG nova.compute.manager [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Refreshing instance network info cache due to event network-changed-0da6d475-040d-43e8-9139-88652285a5fe. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 727.500989] env[70013]: DEBUG oslo_concurrency.lockutils [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] Acquiring lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.502077] env[70013]: DEBUG oslo_concurrency.lockutils [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] Acquired lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.502077] env[70013]: DEBUG nova.network.neutron [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Refreshing network info cache for port 0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 727.561824] env[70013]: DEBUG nova.network.neutron [-] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.592029] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230590, 'name': CreateVM_Task, 'duration_secs': 0.336103} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.592189] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 727.592544] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.592703] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.593068] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 727.593334] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68dc6c02-ac94-48f6-8895-775ecf3cda1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.599228] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27baee5-44ce-411d-9c9b-ad448e19a622 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.603944] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 727.603944] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b09308-b3d0-e8f8-50bd-59b76ac8ba9f" [ 727.603944] env[70013]: _type = "Task" [ 727.603944] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.613169] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492f4d5d-2b22-4785-8e09-a20b7cabd35d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.621706] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b09308-b3d0-e8f8-50bd-59b76ac8ba9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.652283] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cc4c0d-6908-4614-a35f-b632390300b3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.663860] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ef824e-0526-48ec-bc39-353a7b895b93 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.674518] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230593, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.674705] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230594, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.686020] env[70013]: DEBUG nova.compute.provider_tree [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.691090] env[70013]: DEBUG nova.objects.instance [None req-ea689c5e-7942-4663-a907-0d8e2be01ef7 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lazy-loading 'pci_requests' on Instance uuid 225c02d6-43df-4260-a428-14a6b1ed32cc {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.780687] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230595, 'name': Rename_Task, 'duration_secs': 0.329948} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.781036] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 727.781366] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c6a8b1b-4001-4493-9ab3-b4c6ca979a0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.789322] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 727.789322] env[70013]: value = "task-4230597" [ 727.789322] env[70013]: _type = "Task" [ 727.789322] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.805306] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.805306] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.869171] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230596, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.886685] env[70013]: DEBUG nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 728.065619] env[70013]: INFO nova.compute.manager [-] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Took 1.26 seconds to deallocate network for instance. [ 728.117284] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b09308-b3d0-e8f8-50bd-59b76ac8ba9f, 'name': SearchDatastore_Task, 'duration_secs': 0.030725} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.117597] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.117829] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 728.118118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.118278] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.118455] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 728.118791] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eab85bc1-2923-480e-867f-19563489aa1f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.134017] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 728.134271] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 728.135050] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4510c488-afe9-41a6-bd2e-3c95fa2b0197 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.141574] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 728.141574] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e85d98-17d6-622e-71f5-32bbc87e185d" [ 728.141574] env[70013]: _type = "Task" [ 728.141574] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.151537] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e85d98-17d6-622e-71f5-32bbc87e185d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.164024] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230593, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560639} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.164560] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 728.164895] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.165210] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbe4c172-d4c6-41a5-a43b-0d6d50b0676a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.171426] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230594, 'name': CreateVM_Task, 'duration_secs': 0.697089} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.172017] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 728.172843] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.173044] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.173368] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 728.173627] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dfc2c45-c7e9-4623-80c6-abbdb8fd17e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.178917] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 728.178917] env[70013]: value = "task-4230598" [ 728.178917] env[70013]: _type = "Task" [ 728.178917] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.180514] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 728.180514] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d95020-aab1-bb51-101f-02b9cfda0b17" [ 728.180514] env[70013]: _type = "Task" [ 728.180514] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.193121] env[70013]: DEBUG nova.scheduler.client.report [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 728.205247] env[70013]: DEBUG nova.objects.base [None req-ea689c5e-7942-4663-a907-0d8e2be01ef7 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Object Instance<225c02d6-43df-4260-a428-14a6b1ed32cc> lazy-loaded attributes: flavor,pci_requests {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 728.205594] env[70013]: DEBUG nova.network.neutron [None req-ea689c5e-7942-4663-a907-0d8e2be01ef7 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.208254] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d95020-aab1-bb51-101f-02b9cfda0b17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.208254] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230598, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.300916] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.308086] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea689c5e-7942-4663-a907-0d8e2be01ef7 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "interface-225c02d6-43df-4260-a428-14a6b1ed32cc-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.124s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.365310] env[70013]: DEBUG oslo_vmware.api [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230596, 'name': PowerOnVM_Task, 'duration_secs': 0.737771} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.365613] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 728.365875] env[70013]: INFO nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Took 7.98 seconds to spawn the instance on the hypervisor. [ 728.366102] env[70013]: DEBUG nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 728.367579] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a1aa9b-12e3-47b9-95ac-37f1f8d9b6e1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.427309] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.528613] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "062eff58-95fe-4c9b-a586-7e7434c77adf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.528786] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "062eff58-95fe-4c9b-a586-7e7434c77adf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.576752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.660092] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e85d98-17d6-622e-71f5-32bbc87e185d, 'name': SearchDatastore_Task, 'duration_secs': 0.028095} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.660092] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b946237b-5ea5-4877-a8e0-2080135fff0a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.667995] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 728.667995] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cd3083-32ec-62ac-605e-33f0a7fe4f13" [ 728.667995] env[70013]: _type = "Task" [ 728.667995] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.680656] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cd3083-32ec-62ac-605e-33f0a7fe4f13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.697303] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d95020-aab1-bb51-101f-02b9cfda0b17, 'name': SearchDatastore_Task, 'duration_secs': 0.020867} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.698436] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.698743] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 728.698980] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.699369] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.699369] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 728.700430] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230598, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091813} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.700430] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50950766-a6c7-425a-bc80-4fa0b9098a84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.702805] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 728.702805] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8c293d-51b8-4a7a-8c7f-38d715a6e3b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.719604] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.634s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.720187] env[70013]: DEBUG nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 728.733427] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 728.736070] env[70013]: DEBUG nova.network.neutron [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updated VIF entry in instance network info cache for port 0da6d475-040d-43e8-9139-88652285a5fe. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 728.736405] env[70013]: DEBUG nova.network.neutron [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updating instance_info_cache with network_info: [{"id": "0da6d475-040d-43e8-9139-88652285a5fe", "address": "fa:16:3e:1e:a7:37", "network": {"id": "1c5012af-830c-4001-9013-9e8cf32698bb", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-222945680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "719bc6e70189480dbef8b541a7a63d36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0da6d475-04", "ovs_interfaceid": "0da6d475-040d-43e8-9139-88652285a5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.738256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.693s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.738256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.740141] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.610s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.740340] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.745952] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.090s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.749174] env[70013]: INFO nova.compute.claims [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.752046] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0cd3939-2bb9-42dc-8f69-c24abd1df381 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.769088] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 728.769885] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 728.772639] env[70013]: DEBUG oslo_concurrency.lockutils [req-a078b354-1643-40bc-81b7-0ce29d42006d req-78482b77-8b7c-4747-8084-db34744347cb service nova] Releasing lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.772639] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb3528e4-542a-4bad-baf7-41240955fdb4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.785425] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 728.785425] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5277fc8b-f7ee-492b-277b-86fbf5341657" [ 728.785425] env[70013]: _type = "Task" [ 728.785425] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.794776] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 728.794776] env[70013]: value = "task-4230599" [ 728.794776] env[70013]: _type = "Task" [ 728.794776] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.807869] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5277fc8b-f7ee-492b-277b-86fbf5341657, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.820120] env[70013]: DEBUG oslo_vmware.api [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230597, 'name': PowerOnVM_Task, 'duration_secs': 0.985678} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.823432] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 728.823671] env[70013]: INFO nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Took 10.46 seconds to spawn the instance on the hypervisor. [ 728.823850] env[70013]: DEBUG nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 728.824198] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230599, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.829950] env[70013]: INFO nova.scheduler.client.report [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Deleted allocations for instance e8998cab-533b-46b1-a1ea-5eb5db7d6d99 [ 728.836018] env[70013]: INFO nova.scheduler.client.report [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Deleted allocations for instance 2b91ec52-2d32-4f86-acf2-f74f661f65b5 [ 728.836195] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e36df90-4ce6-4dda-bd3c-05ab4d20defd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.891851] env[70013]: INFO nova.compute.manager [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Took 23.14 seconds to build instance. [ 729.180072] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cd3083-32ec-62ac-605e-33f0a7fe4f13, 'name': SearchDatastore_Task, 'duration_secs': 0.044949} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.180357] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.180621] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 729.180940] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7366b05b-b1fe-4db2-ab00-e65d400f941b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.188787] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 729.188787] env[70013]: value = "task-4230600" [ 729.188787] env[70013]: _type = "Task" [ 729.188787] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.199040] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230600, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.237861] env[70013]: DEBUG nova.compute.utils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 729.238640] env[70013]: DEBUG nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 729.238850] env[70013]: DEBUG nova.network.neutron [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 729.300502] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5277fc8b-f7ee-492b-277b-86fbf5341657, 'name': SearchDatastore_Task, 'duration_secs': 0.018582} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.304724] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbddd16a-d8de-4757-b1da-5be55d8e2bdc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.311099] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 729.311099] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52031b92-65f1-35c5-c986-5cdb5d1ef4ae" [ 729.311099] env[70013]: _type = "Task" [ 729.311099] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.319253] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230599, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.324545] env[70013]: DEBUG nova.policy [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e742fb499644961bdac697b0b170131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0aaeba8fc0884a8f85e6af2a6be00b1a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 729.339136] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52031b92-65f1-35c5-c986-5cdb5d1ef4ae, 'name': SearchDatastore_Task, 'duration_secs': 0.015695} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.339803] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.340944] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5/8ec4b86d-e83b-4b28-ac4b-d990177b0ba5.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 729.341329] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d34c4da-8e4a-4f18-8ca7-fe50a26f497e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.354102] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 729.354102] env[70013]: value = "task-4230601" [ 729.354102] env[70013]: _type = "Task" [ 729.354102] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.354434] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f636c5ef-bc35-4af5-b60c-e389e731f3af tempest-DeleteServersAdminTestJSON-2063474761 tempest-DeleteServersAdminTestJSON-2063474761-project-admin] Lock "e8998cab-533b-46b1-a1ea-5eb5db7d6d99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.084s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.370311] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.371137] env[70013]: INFO nova.compute.manager [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Took 24.87 seconds to build instance. [ 729.380197] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b40fa478-c5de-4929-b9bb-c402080e34a9 tempest-ServerDiagnosticsTest-1436260678 tempest-ServerDiagnosticsTest-1436260678-project-member] Lock "2b91ec52-2d32-4f86-acf2-f74f661f65b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.533s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.396042] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2c99fbc0-e6b2-4597-9010-729571ab2059 tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.668s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.705659] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230600, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.745300] env[70013]: DEBUG nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 729.816971] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230599, 'name': ReconfigVM_Task, 'duration_secs': 0.714732} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.821574] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20/98a0199a-4209-4faf-adf5-7ae33b099d20.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 729.822983] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-709fc7f7-0ae4-4325-bf0a-f7c6933f84bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.833408] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 729.833408] env[70013]: value = "task-4230602" [ 729.833408] env[70013]: _type = "Task" [ 729.833408] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.855741] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230602, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.874651] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab08cefb-030a-48b3-9894-df96e47252f0 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.722s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.875740] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230601, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.899444] env[70013]: DEBUG nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 729.964402] env[70013]: DEBUG nova.network.neutron [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Successfully created port: c4a28327-a59d-44df-9cc4-ee9bcd50c21e {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.208734] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230600, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.768123} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.209127] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 730.209281] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 730.209623] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70f47165-c197-4651-ac92-735d9c00b8f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.215989] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc268473-d063-47c0-adce-6890346fd273 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.230792] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0497c582-0f95-409f-b357-3ed42231d849 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.236039] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 730.236039] env[70013]: value = "task-4230603" [ 730.236039] env[70013]: _type = "Task" [ 730.236039] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.285396] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0025ce-56ad-4bb4-9cd4-66284cb98169 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.292453] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230603, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.304682] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f42cd2d-eb68-43b7-bd28-bf22d4ccc83f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.328072] env[70013]: DEBUG nova.compute.provider_tree [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.345149] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230602, 'name': Rename_Task, 'duration_secs': 0.383172} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.346312] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 730.349020] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c0e1a94-780d-4886-86ac-287081befdad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.356381] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Waiting for the task: (returnval){ [ 730.356381] env[70013]: value = "task-4230604" [ 730.356381] env[70013]: _type = "Task" [ 730.356381] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.375915] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230604, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.380899] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.680501} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.380899] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5/8ec4b86d-e83b-4b28-ac4b-d990177b0ba5.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 730.382185] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 730.382185] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba371c8c-8dc5-4c52-8af7-bc8f89348fd3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.390975] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 730.390975] env[70013]: value = "task-4230605" [ 730.390975] env[70013]: _type = "Task" [ 730.390975] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.402274] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.428333] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.750538] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073789} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.751731] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.751847] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104fcd9c-9694-481d-b91c-9b942069f864 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.780274] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.780274] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5abdfbb-16f0-4205-b7fa-03d24dd3a24c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.804036] env[70013]: DEBUG nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 730.807451] env[70013]: DEBUG nova.compute.manager [req-746537b9-82cb-4974-9d02-c92e0c540e9d req-ef30e630-2038-4b32-9bb6-c7bbae4158e3 service nova] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Received event network-vif-deleted-c9bdfb0a-f6f0-4467-b641-a8c04598910f {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 730.816696] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 730.816696] env[70013]: value = "task-4230606" [ 730.816696] env[70013]: _type = "Task" [ 730.816696] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.830057] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230606, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.833164] env[70013]: DEBUG nova.scheduler.client.report [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 730.850240] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 730.850676] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.850824] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 730.850993] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.851197] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 730.851259] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 730.851482] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 730.851668] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 730.851876] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 730.852093] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 730.852242] env[70013]: DEBUG nova.virt.hardware [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 730.853277] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7e9d85-63f9-40fd-a7c0-5bb20882f9df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.864862] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8931866-0df0-4d31-9bfd-150e1f582ff0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.874126] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230604, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.904651] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230605, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.138266} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.905102] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.905893] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22494f5c-54d3-4f14-9a0b-8ae89c7abf74 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.933992] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5/8ec4b86d-e83b-4b28-ac4b-d990177b0ba5.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.934765] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3accdefb-abe2-4a09-91b4-d400efed0c64 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.956635] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 730.956635] env[70013]: value = "task-4230607" [ 730.956635] env[70013]: _type = "Task" [ 730.956635] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.965657] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230607, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.329511] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.341480] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.342043] env[70013]: DEBUG nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 731.346558] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.536s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.348433] env[70013]: INFO nova.compute.claims [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.384314] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230604, 'name': PowerOnVM_Task} progress is 81%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.399947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "225c02d6-43df-4260-a428-14a6b1ed32cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.399947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.399947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "225c02d6-43df-4260-a428-14a6b1ed32cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.399947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.400282] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.401303] env[70013]: INFO nova.compute.manager [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Terminating instance [ 731.470952] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230607, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.830108] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230606, 'name': ReconfigVM_Task, 'duration_secs': 0.779806} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.831133] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Reconfigured VM instance instance-00000008 to attach disk [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193/cf8be667-da84-466b-9b7b-3bdfda6a8193.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.831133] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9b5c182-becd-407c-8811-4db0459107cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.843204] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 731.843204] env[70013]: value = "task-4230608" [ 731.843204] env[70013]: _type = "Task" [ 731.843204] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.856011] env[70013]: DEBUG nova.compute.utils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 731.857594] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230608, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.858165] env[70013]: DEBUG nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 731.858348] env[70013]: DEBUG nova.network.neutron [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 731.882331] env[70013]: DEBUG oslo_vmware.api [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Task: {'id': task-4230604, 'name': PowerOnVM_Task, 'duration_secs': 1.258856} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.883533] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 731.883745] env[70013]: DEBUG nova.compute.manager [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 731.884550] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a558cee7-f081-49a8-b8c3-8690eb870294 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.906783] env[70013]: DEBUG nova.compute.manager [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 731.906783] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.909131] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd54cb8-fdb2-4150-9a2d-d5525cc6913c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.920693] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 731.920693] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fe4e746-4d85-471b-a68e-ca08e6d4309f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.933040] env[70013]: DEBUG oslo_vmware.api [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 731.933040] env[70013]: value = "task-4230609" [ 731.933040] env[70013]: _type = "Task" [ 731.933040] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.944550] env[70013]: DEBUG oslo_vmware.api [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230609, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.971744] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230607, 'name': ReconfigVM_Task, 'duration_secs': 0.641663} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.973375] env[70013]: DEBUG nova.policy [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3bcb0d0bd1e44eb960d676070189f27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3d42e73dc0a4a06a2022d8b54e13e12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 731.976982] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5/8ec4b86d-e83b-4b28-ac4b-d990177b0ba5.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.978261] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b94cc0ba-2e01-4473-8e8d-2b76aa97bc61 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.991535] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 731.991535] env[70013]: value = "task-4230610" [ 731.991535] env[70013]: _type = "Task" [ 731.991535] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.002685] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230610, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.348774] env[70013]: DEBUG nova.network.neutron [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Successfully updated port: c4a28327-a59d-44df-9cc4-ee9bcd50c21e {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.361299] env[70013]: DEBUG nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 732.365711] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230608, 'name': Rename_Task, 'duration_secs': 0.245011} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.367161] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 732.367161] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80f90759-9655-4aed-853e-424088262d9c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.380297] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 732.380297] env[70013]: value = "task-4230611" [ 732.380297] env[70013]: _type = "Task" [ 732.380297] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.393374] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.407606] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.447891] env[70013]: DEBUG oslo_vmware.api [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230609, 'name': PowerOffVM_Task, 'duration_secs': 0.284835} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.448208] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.448384] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.448675] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7be88ad8-fa27-4320-b2ff-c79188c115e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.490745] env[70013]: DEBUG nova.network.neutron [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Successfully created port: 4f1ea900-8772-4876-82d4-0c75f8ab8e22 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.509957] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230610, 'name': Rename_Task, 'duration_secs': 0.205427} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.510119] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 732.513113] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e81d3fb-2d72-4a4d-afc9-e12befec1a43 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.523027] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 732.523027] env[70013]: value = "task-4230613" [ 732.523027] env[70013]: _type = "Task" [ 732.523027] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.538102] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.541308] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.541308] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.541308] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Deleting the datastore file [datastore2] 225c02d6-43df-4260-a428-14a6b1ed32cc {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.541308] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65ab8c49-bb41-466a-940c-3d349110b4fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.549750] env[70013]: DEBUG oslo_vmware.api [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for the task: (returnval){ [ 732.549750] env[70013]: value = "task-4230614" [ 732.549750] env[70013]: _type = "Task" [ 732.549750] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.561925] env[70013]: DEBUG oslo_vmware.api [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.662446] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.662784] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.856327] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.856327] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquired lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 732.856327] env[70013]: DEBUG nova.network.neutron [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.898629] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230611, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.938113] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2f1cdf-43f7-4a80-a0b8-444cf41c52c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.948619] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4367cf29-518b-4c80-916e-3699b0dcdc20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.991702] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a97606-2adb-4e1c-a0d1-889555e76af7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.996953] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "63c3b57c-022a-4eee-b215-6dd89142e659" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.997301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.997405] env[70013]: DEBUG nova.compute.manager [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 732.998481] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a3516c-a58a-49d1-a46c-ef22d9110701 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.005773] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e99262-cb0c-49ed-8bbf-21c57cc0f06d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.012142] env[70013]: DEBUG nova.compute.manager [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 733.012845] env[70013]: DEBUG nova.objects.instance [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lazy-loading 'flavor' on Instance uuid 63c3b57c-022a-4eee-b215-6dd89142e659 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 733.027451] env[70013]: DEBUG nova.compute.provider_tree [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.039736] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230613, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.063160] env[70013]: DEBUG oslo_vmware.api [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Task: {'id': task-4230614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249254} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.063505] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 733.063730] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 733.063974] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.064650] env[70013]: INFO nova.compute.manager [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Took 1.16 seconds to destroy the instance on the hypervisor. [ 733.064650] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 733.064909] env[70013]: DEBUG nova.compute.manager [-] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 733.065078] env[70013]: DEBUG nova.network.neutron [-] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.136748] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.137030] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.166035] env[70013]: DEBUG nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 733.377453] env[70013]: DEBUG nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 733.402603] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230611, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.413918] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 733.414692] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.415178] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 733.415693] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.416150] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 733.417230] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 733.417230] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 733.417230] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 733.417412] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 733.419114] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 733.419114] env[70013]: DEBUG nova.virt.hardware [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 733.419651] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd78a917-ffe9-46a1-84b7-51b26f29bd7b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.428181] env[70013]: DEBUG nova.network.neutron [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.435406] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bac237-3693-4a32-a841-c24bc8efb5e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.534561] env[70013]: DEBUG nova.scheduler.client.report [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.550292] env[70013]: DEBUG oslo_vmware.api [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230613, 'name': PowerOnVM_Task, 'duration_secs': 0.818953} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.550654] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 733.552184] env[70013]: INFO nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Took 10.56 seconds to spawn the instance on the hypervisor. [ 733.552463] env[70013]: DEBUG nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 733.553366] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f6bf8e-fd21-41e0-8a61-b0a90d4206ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.708299] env[70013]: DEBUG nova.network.neutron [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Updating instance_info_cache with network_info: [{"id": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "address": "fa:16:3e:b5:79:fd", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4a28327-a5", "ovs_interfaceid": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.716179] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.824765] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.825106] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.904532] env[70013]: DEBUG oslo_vmware.api [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230611, 'name': PowerOnVM_Task, 'duration_secs': 1.050176} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.905304] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 733.905304] env[70013]: DEBUG nova.compute.manager [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 733.907244] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a6a887-3e05-439d-bf8c-d94e9aea4f8e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.021591] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 734.024088] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5368bb23-a2d3-4c9b-871a-0081ef5f22b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.032667] env[70013]: DEBUG oslo_vmware.api [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 734.032667] env[70013]: value = "task-4230615" [ 734.032667] env[70013]: _type = "Task" [ 734.032667] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.043831] env[70013]: DEBUG oslo_vmware.api [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.045783] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.700s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.046326] env[70013]: DEBUG nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 734.049775] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.668s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.053025] env[70013]: DEBUG nova.objects.instance [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 734.081033] env[70013]: INFO nova.compute.manager [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Took 23.02 seconds to build instance. [ 734.200846] env[70013]: DEBUG nova.network.neutron [-] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.216823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Releasing lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 734.216823] env[70013]: DEBUG nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Instance network_info: |[{"id": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "address": "fa:16:3e:b5:79:fd", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4a28327-a5", "ovs_interfaceid": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 734.216964] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:79:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4a28327-a59d-44df-9cc4-ee9bcd50c21e', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 734.226405] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Creating folder: Project (0aaeba8fc0884a8f85e6af2a6be00b1a). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 734.227913] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "64308d7c-b63b-4fa6-b235-2f0586a888cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.228530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.229981] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "64308d7c-b63b-4fa6-b235-2f0586a888cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.229981] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.229981] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.233115] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f178530a-9181-46ee-a0da-b5f2d4798b58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.238766] env[70013]: INFO nova.compute.manager [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Terminating instance [ 734.256376] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Created folder: Project (0aaeba8fc0884a8f85e6af2a6be00b1a) in parent group-v836999. [ 734.256741] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Creating folder: Instances. Parent ref: group-v837042. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 734.258214] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-051cfebd-2138-4dcd-9724-2e6b6f83c915 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.277849] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Created folder: Instances in parent group-v837042. [ 734.278276] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 734.279831] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 734.280281] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4cffa848-a0c9-4a83-aaf1-7019f7ceac2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.315767] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 734.315767] env[70013]: value = "task-4230618" [ 734.315767] env[70013]: _type = "Task" [ 734.315767] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.333861] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230618, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.336295] env[70013]: DEBUG nova.network.neutron [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Successfully updated port: 4f1ea900-8772-4876-82d4-0c75f8ab8e22 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 734.426975] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.547681] env[70013]: DEBUG oslo_vmware.api [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230615, 'name': PowerOffVM_Task, 'duration_secs': 0.289999} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.548040] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.548273] env[70013]: DEBUG nova.compute.manager [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 734.549168] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339acd4b-bb99-4854-aade-c16cc48ba121 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.559450] env[70013]: DEBUG nova.compute.utils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 734.564016] env[70013]: DEBUG nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 734.564016] env[70013]: DEBUG nova.network.neutron [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 734.585777] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c472705b-a661-4ae6-a77a-91d8a504d402 tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.375s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.623338] env[70013]: DEBUG nova.policy [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c840837f55364b30a94a13030fc757af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3eb984e7011845a1b0b848d8a17c6990', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 734.704966] env[70013]: INFO nova.compute.manager [-] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Took 1.64 seconds to deallocate network for instance. [ 734.745745] env[70013]: DEBUG nova.compute.manager [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 734.745745] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 734.748109] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bfc172-8ace-4ac1-9d07-5671e93037b4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.760674] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 734.761075] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43efa8a8-16d0-48ea-8ad4-cecca1aabd5a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.771080] env[70013]: DEBUG oslo_vmware.api [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 734.771080] env[70013]: value = "task-4230619" [ 734.771080] env[70013]: _type = "Task" [ 734.771080] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.782437] env[70013]: DEBUG nova.compute.manager [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Received event network-changed-a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 734.782654] env[70013]: DEBUG nova.compute.manager [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Refreshing instance network info cache due to event network-changed-a0d11944-d177-42a5-ad2b-22c116396f8a. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 734.783312] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.783555] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.783761] env[70013]: DEBUG nova.network.neutron [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Refreshing network info cache for port a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 734.797696] env[70013]: DEBUG oslo_vmware.api [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.828461] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230618, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.838847] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.839162] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.839415] env[70013]: DEBUG nova.network.neutron [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.062753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6688c8fe-5373-4d31-9151-62abf6481696 tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.062753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.243s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.062753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.065901] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 14.447s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.071151] env[70013]: DEBUG nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 735.074415] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7483a6d4-b402-4621-8ad1-7ff43cb4fee2 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.077s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.092178] env[70013]: DEBUG nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 735.099112] env[70013]: INFO nova.scheduler.client.report [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Deleted allocations for instance 9c5f2422-820e-454b-98be-695e3ac6d1e3 [ 735.219869] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.284517] env[70013]: DEBUG oslo_vmware.api [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230619, 'name': PowerOffVM_Task, 'duration_secs': 0.495909} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.285243] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 735.285243] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 735.285619] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3097968f-1588-4829-92ab-b5b78eea0530 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.330048] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230618, 'name': CreateVM_Task, 'duration_secs': 0.581358} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.330283] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 735.331668] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.331821] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 735.332214] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 735.332868] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb2564c0-95bf-4a53-a0ce-ecf28fa98358 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.335557] env[70013]: DEBUG nova.network.neutron [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Successfully created port: 47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.345070] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 735.345070] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ddb897-f96c-6b3a-f469-0aaa263855f0" [ 735.345070] env[70013]: _type = "Task" [ 735.345070] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.355941] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ddb897-f96c-6b3a-f469-0aaa263855f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.362999] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 735.363369] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 735.363544] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Deleting the datastore file [datastore2] 64308d7c-b63b-4fa6-b235-2f0586a888cf {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 735.363908] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-177d2ce3-cc64-4190-aaae-2ed7bb3d89d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.372774] env[70013]: DEBUG oslo_vmware.api [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for the task: (returnval){ [ 735.372774] env[70013]: value = "task-4230621" [ 735.372774] env[70013]: _type = "Task" [ 735.372774] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.382670] env[70013]: DEBUG oslo_vmware.api [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.431239] env[70013]: DEBUG nova.network.neutron [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.455013] env[70013]: DEBUG nova.compute.manager [req-64392e95-734f-4d96-a1bd-1c5194d44c31 req-161a618a-cf05-40fd-93fe-e624f890ca50 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Received event network-vif-plugged-4f1ea900-8772-4876-82d4-0c75f8ab8e22 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 735.455282] env[70013]: DEBUG oslo_concurrency.lockutils [req-64392e95-734f-4d96-a1bd-1c5194d44c31 req-161a618a-cf05-40fd-93fe-e624f890ca50 service nova] Acquiring lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.455455] env[70013]: DEBUG oslo_concurrency.lockutils [req-64392e95-734f-4d96-a1bd-1c5194d44c31 req-161a618a-cf05-40fd-93fe-e624f890ca50 service nova] Lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.455623] env[70013]: DEBUG oslo_concurrency.lockutils [req-64392e95-734f-4d96-a1bd-1c5194d44c31 req-161a618a-cf05-40fd-93fe-e624f890ca50 service nova] Lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.455790] env[70013]: DEBUG nova.compute.manager [req-64392e95-734f-4d96-a1bd-1c5194d44c31 req-161a618a-cf05-40fd-93fe-e624f890ca50 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] No waiting events found dispatching network-vif-plugged-4f1ea900-8772-4876-82d4-0c75f8ab8e22 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 735.455953] env[70013]: WARNING nova.compute.manager [req-64392e95-734f-4d96-a1bd-1c5194d44c31 req-161a618a-cf05-40fd-93fe-e624f890ca50 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Received unexpected event network-vif-plugged-4f1ea900-8772-4876-82d4-0c75f8ab8e22 for instance with vm_state building and task_state spawning. [ 735.573855] env[70013]: INFO nova.compute.claims [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.609778] env[70013]: DEBUG oslo_concurrency.lockutils [None req-001cad16-da0e-4381-8e31-0d436f866fba tempest-ImagesNegativeTestJSON-742815310 tempest-ImagesNegativeTestJSON-742815310-project-member] Lock "9c5f2422-820e-454b-98be-695e3ac6d1e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.628s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.638526] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.860714] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ddb897-f96c-6b3a-f469-0aaa263855f0, 'name': SearchDatastore_Task, 'duration_secs': 0.014174} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.861884] env[70013]: DEBUG nova.network.neutron [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Updating instance_info_cache with network_info: [{"id": "4f1ea900-8772-4876-82d4-0c75f8ab8e22", "address": "fa:16:3e:21:bb:46", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f1ea900-87", "ovs_interfaceid": "4f1ea900-8772-4876-82d4-0c75f8ab8e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.865424] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 735.865701] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 735.865957] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.866120] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 735.866367] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 735.867518] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99ef4a6b-1da9-4070-acc2-5100132c47c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.880707] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 735.880707] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 735.882513] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c691a6c-fc24-455e-bcfc-e1ff87d214ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.889412] env[70013]: DEBUG oslo_vmware.api [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Task: {'id': task-4230621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.496691} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.891140] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 735.891140] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 735.891335] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.891464] env[70013]: INFO nova.compute.manager [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Took 1.15 seconds to destroy the instance on the hypervisor. [ 735.891706] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 735.892235] env[70013]: DEBUG nova.compute.manager [-] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 735.892338] env[70013]: DEBUG nova.network.neutron [-] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.897575] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 735.897575] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c42694-f159-bd6b-79ef-603a877560bc" [ 735.897575] env[70013]: _type = "Task" [ 735.897575] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.907386] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c42694-f159-bd6b-79ef-603a877560bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.060581] env[70013]: DEBUG nova.network.neutron [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updated VIF entry in instance network info cache for port a0d11944-d177-42a5-ad2b-22c116396f8a. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 736.061070] env[70013]: DEBUG nova.network.neutron [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.090920] env[70013]: INFO nova.compute.resource_tracker [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating resource usage from migration 2b86447b-3ba9-457a-94d4-7269f0b254ab [ 736.096625] env[70013]: DEBUG nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 736.138669] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 736.138669] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.138872] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 736.138916] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.139591] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 736.139591] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 736.139591] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 736.139591] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 736.139868] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 736.139868] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 736.140137] env[70013]: DEBUG nova.virt.hardware [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 736.141959] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69994698-3dd4-4b28-9690-3f3daecda221 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.157488] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07baf19c-1801-494e-b3f9-d061ed6b150e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.244655] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "98a0199a-4209-4faf-adf5-7ae33b099d20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.245680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.245680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "98a0199a-4209-4faf-adf5-7ae33b099d20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.245680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.245680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.247677] env[70013]: INFO nova.compute.manager [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Terminating instance [ 736.376175] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.376610] env[70013]: DEBUG nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance network_info: |[{"id": "4f1ea900-8772-4876-82d4-0c75f8ab8e22", "address": "fa:16:3e:21:bb:46", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f1ea900-87", "ovs_interfaceid": "4f1ea900-8772-4876-82d4-0c75f8ab8e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 736.378168] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:bb:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f1ea900-8772-4876-82d4-0c75f8ab8e22', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.386815] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating folder: Project (c3d42e73dc0a4a06a2022d8b54e13e12). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.387399] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-50c0bc22-f251-48d4-be56-d4540a4b84e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.405018] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Created folder: Project (c3d42e73dc0a4a06a2022d8b54e13e12) in parent group-v836999. [ 736.405326] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating folder: Instances. Parent ref: group-v837045. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.406195] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0e86d2b-668a-4377-942e-8e3c21523bec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.421993] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c42694-f159-bd6b-79ef-603a877560bc, 'name': SearchDatastore_Task, 'duration_secs': 0.013313} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.424703] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Created folder: Instances in parent group-v837045. [ 736.428029] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 736.428029] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44137bd7-c095-41ad-ab51-756dcc441b4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.428029] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 736.428219] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-490ed93c-debc-4b51-875a-0262c20353a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.452278] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 736.452278] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e4a93c-8b76-54a1-1926-eb5a194152da" [ 736.452278] env[70013]: _type = "Task" [ 736.452278] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.456997] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 736.456997] env[70013]: value = "task-4230624" [ 736.456997] env[70013]: _type = "Task" [ 736.456997] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.466756] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e4a93c-8b76-54a1-1926-eb5a194152da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.472735] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230624, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.533197] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51d7231-83dd-45eb-bd67-3a483e28497c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.542965] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5acee8-1db6-43c1-ba3b-f2803176bbf3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.584664] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.584926] env[70013]: DEBUG nova.compute.manager [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Received event network-vif-plugged-c4a28327-a59d-44df-9cc4-ee9bcd50c21e {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 736.585150] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Acquiring lock "fc0b7f87-8a36-440e-b158-30b9d9920572-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.585357] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.585515] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.585686] env[70013]: DEBUG nova.compute.manager [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] No waiting events found dispatching network-vif-plugged-c4a28327-a59d-44df-9cc4-ee9bcd50c21e {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 736.585855] env[70013]: WARNING nova.compute.manager [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Received unexpected event network-vif-plugged-c4a28327-a59d-44df-9cc4-ee9bcd50c21e for instance with vm_state building and task_state spawning. [ 736.586080] env[70013]: DEBUG nova.compute.manager [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Received event network-changed-c4a28327-a59d-44df-9cc4-ee9bcd50c21e {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 736.586293] env[70013]: DEBUG nova.compute.manager [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Refreshing instance network info cache due to event network-changed-c4a28327-a59d-44df-9cc4-ee9bcd50c21e. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 736.586491] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Acquiring lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.586644] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Acquired lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.586832] env[70013]: DEBUG nova.network.neutron [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Refreshing network info cache for port c4a28327-a59d-44df-9cc4-ee9bcd50c21e {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.588936] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717ec6d4-0b13-4d05-a356-16b3d126fa12 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.598184] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7db6f8-11c9-4aaf-b5db-240d0c335c75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.617398] env[70013]: DEBUG nova.compute.provider_tree [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.756261] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "refresh_cache-98a0199a-4209-4faf-adf5-7ae33b099d20" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.757504] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquired lock "refresh_cache-98a0199a-4209-4faf-adf5-7ae33b099d20" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.757504] env[70013]: DEBUG nova.network.neutron [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.966558] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e4a93c-8b76-54a1-1926-eb5a194152da, 'name': SearchDatastore_Task, 'duration_secs': 0.027485} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.967624] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.968082] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] fc0b7f87-8a36-440e-b158-30b9d9920572/fc0b7f87-8a36-440e-b158-30b9d9920572.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 736.968214] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c459f08-2336-4a66-942e-08b6e208d833 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.974664] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230624, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.981640] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 736.981640] env[70013]: value = "task-4230625" [ 736.981640] env[70013]: _type = "Task" [ 736.981640] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.990974] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.121549] env[70013]: DEBUG nova.scheduler.client.report [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 737.311250] env[70013]: DEBUG nova.network.neutron [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.314374] env[70013]: DEBUG nova.network.neutron [-] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.470380] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230624, 'name': CreateVM_Task, 'duration_secs': 0.683138} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.471648] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 737.472455] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.472752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.473467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 737.474443] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-600db034-47a6-42b5-9955-e05ca3cf2dee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.482055] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 737.482055] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f01735-21fa-1efe-db86-c6ee31142592" [ 737.482055] env[70013]: _type = "Task" [ 737.482055] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.486534] env[70013]: DEBUG nova.network.neutron [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.502393] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f01735-21fa-1efe-db86-c6ee31142592, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.506243] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230625, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.627738] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.562s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 737.627940] env[70013]: INFO nova.compute.manager [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Migrating [ 737.628285] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.628446] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "compute-rpcapi-router" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.630522] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.406s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 737.632231] env[70013]: INFO nova.compute.claims [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.636728] env[70013]: DEBUG nova.network.neutron [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Updated VIF entry in instance network info cache for port c4a28327-a59d-44df-9cc4-ee9bcd50c21e. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 737.637642] env[70013]: DEBUG nova.network.neutron [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Updating instance_info_cache with network_info: [{"id": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "address": "fa:16:3e:b5:79:fd", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4a28327-a5", "ovs_interfaceid": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.639641] env[70013]: INFO nova.compute.rpcapi [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 737.639992] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "compute-rpcapi-router" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.656424] env[70013]: DEBUG oslo_concurrency.lockutils [req-8c68c776-fe1a-4ab4-93a5-785e93a45a8f req-37e830b9-2a5f-4d06-8afe-f7034431e9f0 service nova] Releasing lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.813019] env[70013]: DEBUG nova.network.neutron [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Successfully updated port: 47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 737.821332] env[70013]: INFO nova.compute.manager [-] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Took 1.93 seconds to deallocate network for instance. [ 737.991747] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Releasing lock "refresh_cache-98a0199a-4209-4faf-adf5-7ae33b099d20" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.991747] env[70013]: DEBUG nova.compute.manager [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 737.992062] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.996489] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ae1e97-e406-4b75-82ef-df2d23efd57c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.006338] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230625, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.87453} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.008925] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] fc0b7f87-8a36-440e-b158-30b9d9920572/fc0b7f87-8a36-440e-b158-30b9d9920572.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 738.008925] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 738.008925] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c042b55c-b412-49e3-9f50-aacb16eb975c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.016873] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f01735-21fa-1efe-db86-c6ee31142592, 'name': SearchDatastore_Task, 'duration_secs': 0.0727} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.017536] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "cf8be667-da84-466b-9b7b-3bdfda6a8193" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.017705] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "cf8be667-da84-466b-9b7b-3bdfda6a8193" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.018135] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "cf8be667-da84-466b-9b7b-3bdfda6a8193-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.018135] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "cf8be667-da84-466b-9b7b-3bdfda6a8193-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.019382] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "cf8be667-da84-466b-9b7b-3bdfda6a8193-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.020421] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 738.020755] env[70013]: INFO nova.compute.manager [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Terminating instance [ 738.022568] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 738.022820] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.023570] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.023570] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.023570] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.024260] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63da1fc2-4fb1-4138-8f64-7c61bd292ccf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.027494] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28e644b1-fba6-4aea-b27f-951713d7c44f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.031355] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 738.031355] env[70013]: value = "task-4230626" [ 738.031355] env[70013]: _type = "Task" [ 738.031355] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.038120] env[70013]: DEBUG oslo_vmware.api [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 738.038120] env[70013]: value = "task-4230627" [ 738.038120] env[70013]: _type = "Task" [ 738.038120] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.047789] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.048383] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 738.049338] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.050106] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8c0e70a-7343-4928-a625-a547779591a2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.062591] env[70013]: DEBUG oslo_vmware.api [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.067779] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 738.067779] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]520efd66-d3df-b4ec-4630-5a52093234c9" [ 738.067779] env[70013]: _type = "Task" [ 738.067779] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.078202] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520efd66-d3df-b4ec-4630-5a52093234c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.164909] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.165399] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.165399] env[70013]: DEBUG nova.network.neutron [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.318116] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.318284] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquired lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.318457] env[70013]: DEBUG nova.network.neutron [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.331716] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.368174] env[70013]: DEBUG nova.compute.manager [req-e285dcd3-f706-4d9f-acad-b3b886e33566 req-0c01c181-34ae-47c9-830c-7bc88e6ce00d service nova] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Received event network-vif-deleted-aa55189e-9d4e-4fc1-9143-9b0ade9887b7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 738.531257] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "refresh_cache-cf8be667-da84-466b-9b7b-3bdfda6a8193" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.531467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquired lock "refresh_cache-cf8be667-da84-466b-9b7b-3bdfda6a8193" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.531683] env[70013]: DEBUG nova.network.neutron [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.551241] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077395} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.552465] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.553451] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8451302c-a009-417f-9c08-d8bfaad1bf36 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.560561] env[70013]: DEBUG oslo_vmware.api [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230627, 'name': PowerOffVM_Task, 'duration_secs': 0.162717} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.561272] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 738.561445] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 738.561698] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f970f2d-4fad-426c-a8d4-02aa58fc57e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.584897] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Reconfiguring VM instance instance-0000000e to attach disk [datastore1] fc0b7f87-8a36-440e-b158-30b9d9920572/fc0b7f87-8a36-440e-b158-30b9d9920572.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.590460] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-459b5480-d3af-475e-b031-f903d00a5836 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.607190] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 738.607416] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 738.607591] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleting the datastore file [datastore2] 98a0199a-4209-4faf-adf5-7ae33b099d20 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 738.608553] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3aa6c272-bdb8-4c63-be63-c693ab5f9181 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.616324] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520efd66-d3df-b4ec-4630-5a52093234c9, 'name': SearchDatastore_Task, 'duration_secs': 0.011561} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.619611] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 738.619611] env[70013]: value = "task-4230629" [ 738.619611] env[70013]: _type = "Task" [ 738.619611] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.620010] env[70013]: DEBUG oslo_vmware.api [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for the task: (returnval){ [ 738.620010] env[70013]: value = "task-4230630" [ 738.620010] env[70013]: _type = "Task" [ 738.620010] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.620421] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-505f3dcc-237f-4301-a442-fd6f012a0040 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.640378] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 738.640378] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]524c19f5-c9dc-9d5a-d217-e6524582de77" [ 738.640378] env[70013]: _type = "Task" [ 738.640378] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.640378] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230629, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.642142] env[70013]: DEBUG oslo_vmware.api [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.652300] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524c19f5-c9dc-9d5a-d217-e6524582de77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.840973] env[70013]: DEBUG nova.compute.manager [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Received event network-changed-4f1ea900-8772-4876-82d4-0c75f8ab8e22 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 738.841268] env[70013]: DEBUG nova.compute.manager [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Refreshing instance network info cache due to event network-changed-4f1ea900-8772-4876-82d4-0c75f8ab8e22. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 738.841541] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] Acquiring lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.841733] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] Acquired lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.841930] env[70013]: DEBUG nova.network.neutron [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Refreshing network info cache for port 4f1ea900-8772-4876-82d4-0c75f8ab8e22 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 738.892817] env[70013]: DEBUG nova.network.neutron [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.917884] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "67b90fec-a6fd-471d-a425-82bc63de572a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.917884] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "67b90fec-a6fd-471d-a425-82bc63de572a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.084125] env[70013]: DEBUG nova.network.neutron [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.151883] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230629, 'name': ReconfigVM_Task, 'duration_secs': 0.323318} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.151883] env[70013]: DEBUG oslo_vmware.api [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Task: {'id': task-4230630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388513} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.152933] env[70013]: DEBUG nova.network.neutron [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Updating instance_info_cache with network_info: [{"id": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "address": "fa:16:3e:03:9c:38", "network": {"id": "6055f3ca-7760-404c-8618-7f0e0cdc4237", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1361122985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eb984e7011845a1b0b848d8a17c6990", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47970e5c-2d", "ovs_interfaceid": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.159189] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Reconfigured VM instance instance-0000000e to attach disk [datastore1] fc0b7f87-8a36-440e-b158-30b9d9920572/fc0b7f87-8a36-440e-b158-30b9d9920572.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.160030] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 739.160289] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 739.160641] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.160704] env[70013]: INFO nova.compute.manager [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Took 1.17 seconds to destroy the instance on the hypervisor. [ 739.160971] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 739.167268] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba0874c1-2fd7-435e-9b95-290c7c4065dc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.168527] env[70013]: DEBUG nova.compute.manager [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 739.168711] env[70013]: DEBUG nova.network.neutron [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.185357] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524c19f5-c9dc-9d5a-d217-e6524582de77, 'name': SearchDatastore_Task, 'duration_secs': 0.03335} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.186478] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7188dc4-ed3f-46b1-9c1f-4a1994d2df4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.194210] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.194210] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] fc473d40-b57b-437e-9511-58a0ba700a69/fc473d40-b57b-437e-9511-58a0ba700a69.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 739.194210] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 739.194210] env[70013]: value = "task-4230631" [ 739.194210] env[70013]: _type = "Task" [ 739.194210] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.194210] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6362a0e9-91f9-4bf5-936e-f1fc76edbd55 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.203025] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03bec93-78ea-4d47-83ac-b98d0136aaeb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.210368] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 739.210368] env[70013]: value = "task-4230632" [ 739.210368] env[70013]: _type = "Task" [ 739.210368] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.210996] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230631, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.245983] env[70013]: DEBUG nova.network.neutron [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.253871] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb1faab-3318-4c0d-a716-69e47478a533 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.261046] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "cfe1478c-53f2-4df8-99a8-ab23209f5306" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 739.261290] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.262672] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.270702] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90f91c9-5674-4eb2-a0b1-e6d979bf1c63 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.287650] env[70013]: DEBUG nova.compute.provider_tree [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.290814] env[70013]: DEBUG nova.network.neutron [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.667778] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Releasing lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.668552] env[70013]: DEBUG nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Instance network_info: |[{"id": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "address": "fa:16:3e:03:9c:38", "network": {"id": "6055f3ca-7760-404c-8618-7f0e0cdc4237", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1361122985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eb984e7011845a1b0b848d8a17c6990", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47970e5c-2d", "ovs_interfaceid": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 739.669382] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:9c:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73e099e8-2acc-4628-a60d-0b4afa46b39d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47970e5c-2dab-40b7-abcc-f2bc90bcfad7', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.677190] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Creating folder: Project (3eb984e7011845a1b0b848d8a17c6990). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 739.680104] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c394d028-cf85-4768-a84b-7ba629411aea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.686854] env[70013]: DEBUG nova.network.neutron [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance_info_cache with network_info: [{"id": "b402fce4-7392-4139-8de5-7620820e0814", "address": "fa:16:3e:44:ef:b6", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb402fce4-73", "ovs_interfaceid": "b402fce4-7392-4139-8de5-7620820e0814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.693330] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Created folder: Project (3eb984e7011845a1b0b848d8a17c6990) in parent group-v836999. [ 739.693467] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Creating folder: Instances. Parent ref: group-v837048. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 739.693902] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35612a96-4d62-4d11-8a06-071c8478e0da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.708661] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230631, 'name': Rename_Task, 'duration_secs': 0.161415} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.713232] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 739.713969] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Created folder: Instances in parent group-v837048. [ 739.714204] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 739.714877] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-747adfc3-309c-4e95-ae83-6be19ab77f23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.719616] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 739.720476] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e1846f1-4b18-4d05-883e-f93c9345e017 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.743164] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230632, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.746697] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 739.746697] env[70013]: value = "task-4230635" [ 739.746697] env[70013]: _type = "Task" [ 739.746697] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.746697] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 739.746697] env[70013]: value = "task-4230636" [ 739.746697] env[70013]: _type = "Task" [ 739.746697] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.753487] env[70013]: DEBUG nova.network.neutron [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.762270] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230636, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.765987] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.794884] env[70013]: DEBUG nova.scheduler.client.report [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.799666] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Releasing lock "refresh_cache-cf8be667-da84-466b-9b7b-3bdfda6a8193" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.799877] env[70013]: DEBUG nova.compute.manager [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 739.800196] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.801976] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3e8a93-b8fb-4c9b-ab4a-b7c631d10480 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.814914] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 739.814914] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e6da4ec-b2fd-4024-885c-f6275c599ae3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.825924] env[70013]: DEBUG oslo_vmware.api [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 739.825924] env[70013]: value = "task-4230637" [ 739.825924] env[70013]: _type = "Task" [ 739.825924] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.842302] env[70013]: DEBUG oslo_vmware.api [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230637, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.869349] env[70013]: DEBUG nova.network.neutron [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Updated VIF entry in instance network info cache for port 4f1ea900-8772-4876-82d4-0c75f8ab8e22. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 739.870632] env[70013]: DEBUG nova.network.neutron [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Updating instance_info_cache with network_info: [{"id": "4f1ea900-8772-4876-82d4-0c75f8ab8e22", "address": "fa:16:3e:21:bb:46", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f1ea900-87", "ovs_interfaceid": "4f1ea900-8772-4876-82d4-0c75f8ab8e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.185763] env[70013]: DEBUG nova.compute.manager [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 740.187181] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16b73bd-22f0-4bc8-bbc9-24315b904e20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.194469] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.229484] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.642516} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.229823] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] fc473d40-b57b-437e-9511-58a0ba700a69/fc473d40-b57b-437e-9511-58a0ba700a69.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 740.230056] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 740.230325] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-421523ab-89b7-4ffd-abca-d08c81056cd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.239024] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 740.239024] env[70013]: value = "task-4230638" [ 740.239024] env[70013]: _type = "Task" [ 740.239024] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.252151] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230638, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.264584] env[70013]: INFO nova.compute.manager [-] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Took 1.10 seconds to deallocate network for instance. [ 740.291910] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230636, 'name': CreateVM_Task, 'duration_secs': 0.424509} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.292389] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230635, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.293297] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 740.294821] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.295009] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.296335] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 740.296740] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9a1a251-bb4a-471f-b387-51b2472e277c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.304701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.675s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.305407] env[70013]: DEBUG nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 740.313032] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 740.313032] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b86998-53bd-bc95-e3a7-c51e1a48e5ee" [ 740.313032] env[70013]: _type = "Task" [ 740.313032] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.313032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.452s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.313032] env[70013]: INFO nova.compute.claims [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.331677] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b86998-53bd-bc95-e3a7-c51e1a48e5ee, 'name': SearchDatastore_Task, 'duration_secs': 0.014376} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.335534] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.338432] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.338432] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.338432] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.338432] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.338432] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c3f4c5a-25aa-46bb-a9d0-48bafac57520 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.346970] env[70013]: DEBUG oslo_vmware.api [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230637, 'name': PowerOffVM_Task, 'duration_secs': 0.181294} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.347354] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 740.347597] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 740.347937] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-209cdbbb-fcb3-4a1d-9bfd-3ced1ea7ecfb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.351894] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.351894] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 740.354326] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0be382bb-8941-4bc1-a993-eddf7006e255 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.362409] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 740.362409] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f05df9-1d5f-3121-5676-f156bed3ae9d" [ 740.362409] env[70013]: _type = "Task" [ 740.362409] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.371655] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f05df9-1d5f-3121-5676-f156bed3ae9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.373768] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f3de151-185e-4800-800a-b10dbadf7c57 req-a268e274-e898-4550-a500-a1ca89d75dc1 service nova] Releasing lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.381999] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 740.382309] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 740.382508] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Deleting the datastore file [datastore1] cf8be667-da84-466b-9b7b-3bdfda6a8193 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.382794] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57f47bc2-d7c5-4379-be81-ab8bc0f51e9a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.392177] env[70013]: DEBUG oslo_vmware.api [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for the task: (returnval){ [ 740.392177] env[70013]: value = "task-4230640" [ 740.392177] env[70013]: _type = "Task" [ 740.392177] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.402856] env[70013]: DEBUG oslo_vmware.api [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.710764] env[70013]: INFO nova.compute.manager [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] instance snapshotting [ 740.710992] env[70013]: WARNING nova.compute.manager [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 740.714833] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0545e0ed-0326-4d10-a738-3eeac3efc401 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.736780] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81279f3-d72e-479d-999a-4fe9121769ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.755704] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230638, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086272} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.759426] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.760564] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997edf96-7d22-4b45-be8d-9dbab1501512 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.772056] env[70013]: DEBUG oslo_vmware.api [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230635, 'name': PowerOnVM_Task, 'duration_secs': 0.600176} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.781169] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 740.781169] env[70013]: INFO nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Took 9.98 seconds to spawn the instance on the hypervisor. [ 740.781464] env[70013]: DEBUG nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 740.794612] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Reconfiguring VM instance instance-0000000f to attach disk [datastore1] fc473d40-b57b-437e-9511-58a0ba700a69/fc473d40-b57b-437e-9511-58a0ba700a69.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.795633] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cad101-5728-4385-be86-38a23fd082a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.800774] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 740.801036] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-513a0d51-7172-44d0-ade4-b3ac010c1ae7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.817384] env[70013]: DEBUG nova.compute.utils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 740.822055] env[70013]: DEBUG nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 740.822594] env[70013]: DEBUG nova.network.neutron [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.833345] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 740.833345] env[70013]: value = "task-4230641" [ 740.833345] env[70013]: _type = "Task" [ 740.833345] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.847467] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230641, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.871601] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f05df9-1d5f-3121-5676-f156bed3ae9d, 'name': SearchDatastore_Task, 'duration_secs': 0.013223} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.873160] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70ddee07-698f-4c1b-8357-102880ad7d12 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.881911] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 740.881911] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529f2f1c-f2f4-82ca-e820-c42748c6234d" [ 740.881911] env[70013]: _type = "Task" [ 740.881911] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.897348] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529f2f1c-f2f4-82ca-e820-c42748c6234d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.909578] env[70013]: DEBUG oslo_vmware.api [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Task: {'id': task-4230640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190973} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.910221] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.910420] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 740.910955] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.910955] env[70013]: INFO nova.compute.manager [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Took 1.11 seconds to destroy the instance on the hypervisor. [ 740.911095] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 740.911236] env[70013]: DEBUG nova.compute.manager [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 740.912057] env[70013]: DEBUG nova.network.neutron [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.916825] env[70013]: DEBUG nova.policy [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58b11e5de683402594c8e5bf66566406', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0dd47358d7b5432d93bddf797588a0cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 740.949648] env[70013]: DEBUG nova.network.neutron [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.261848] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 741.262985] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bcb4c87a-ec16-4567-8c55-ef136f5a2ddb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.271403] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 741.271403] env[70013]: value = "task-4230642" [ 741.271403] env[70013]: _type = "Task" [ 741.271403] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.281104] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230642, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.324074] env[70013]: DEBUG nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 741.355451] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230641, 'name': ReconfigVM_Task, 'duration_secs': 0.290768} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.355745] env[70013]: INFO nova.compute.manager [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Took 26.80 seconds to build instance. [ 741.356676] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Reconfigured VM instance instance-0000000f to attach disk [datastore1] fc473d40-b57b-437e-9511-58a0ba700a69/fc473d40-b57b-437e-9511-58a0ba700a69.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 741.360194] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52761ffa-d983-4631-9761-dcb690437fae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.366267] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 741.366267] env[70013]: value = "task-4230643" [ 741.366267] env[70013]: _type = "Task" [ 741.366267] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.379029] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230643, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.407597] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529f2f1c-f2f4-82ca-e820-c42748c6234d, 'name': SearchDatastore_Task, 'duration_secs': 0.01559} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.408168] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.408168] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25/b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 741.408505] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e0767ce-6d05-4ec1-b6bf-3d040d804148 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.416445] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 741.416445] env[70013]: value = "task-4230644" [ 741.416445] env[70013]: _type = "Task" [ 741.416445] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.427371] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.451683] env[70013]: DEBUG nova.network.neutron [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.497259] env[70013]: DEBUG nova.network.neutron [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Successfully created port: 3e878689-ac09-469d-8490-a73b6e02d403 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.525460] env[70013]: DEBUG nova.compute.manager [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Received event network-vif-deleted-9a809fa3-d1c4-4bff-b415-a9845cbdc5ad {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 741.525744] env[70013]: DEBUG nova.compute.manager [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Received event network-changed-0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 741.525877] env[70013]: DEBUG nova.compute.manager [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Refreshing instance network info cache due to event network-changed-0da6d475-040d-43e8-9139-88652285a5fe. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 741.526603] env[70013]: DEBUG oslo_concurrency.lockutils [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] Acquiring lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.527118] env[70013]: DEBUG oslo_concurrency.lockutils [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] Acquired lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.527118] env[70013]: DEBUG nova.network.neutron [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Refreshing network info cache for port 0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 741.725423] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24db03c8-7b97-4997-a246-cb18eb6f18f2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.755281] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance '3b29c99a-0d55-40b1-a155-199b1ced2146' progress to 0 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 741.786814] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230642, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.859224] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5cd53925-a5a0-42a9-8e16-67f8260b062e tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.342s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.867627] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab357ed-a585-4bce-aa7f-6dcf7da04fb9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.883125] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230643, 'name': Rename_Task, 'duration_secs': 0.165567} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.885857] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 741.887646] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ccaed4b-abf7-4b6a-8989-596c4fb6ef04 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.890556] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6ac876-b5f4-4a7b-bce3-16221a7a6271 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.929539] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2be661-f251-439e-be34-67179060ef50 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.932695] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 741.932695] env[70013]: value = "task-4230645" [ 741.932695] env[70013]: _type = "Task" [ 741.932695] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.944691] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230644, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.945928] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6478b9d7-791b-4e35-ab89-a3f5b300502d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.956128] env[70013]: INFO nova.compute.manager [-] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Took 1.04 seconds to deallocate network for instance. [ 741.956128] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230645, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.970251] env[70013]: DEBUG nova.compute.provider_tree [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.092206] env[70013]: DEBUG nova.compute.manager [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Received event network-vif-plugged-47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 742.092437] env[70013]: DEBUG oslo_concurrency.lockutils [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] Acquiring lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.092642] env[70013]: DEBUG oslo_concurrency.lockutils [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.096187] env[70013]: DEBUG oslo_concurrency.lockutils [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.096187] env[70013]: DEBUG nova.compute.manager [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] No waiting events found dispatching network-vif-plugged-47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 742.096187] env[70013]: WARNING nova.compute.manager [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Received unexpected event network-vif-plugged-47970e5c-2dab-40b7-abcc-f2bc90bcfad7 for instance with vm_state building and task_state spawning. [ 742.096379] env[70013]: DEBUG nova.compute.manager [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Received event network-changed-47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 742.096419] env[70013]: DEBUG nova.compute.manager [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Refreshing instance network info cache due to event network-changed-47970e5c-2dab-40b7-abcc-f2bc90bcfad7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 742.096605] env[70013]: DEBUG oslo_concurrency.lockutils [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] Acquiring lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.096744] env[70013]: DEBUG oslo_concurrency.lockutils [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] Acquired lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.097063] env[70013]: DEBUG nova.network.neutron [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Refreshing network info cache for port 47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 742.266020] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 742.266020] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43ff835c-02bd-4504-8706-593409f400fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.274022] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 742.274022] env[70013]: value = "task-4230646" [ 742.274022] env[70013]: _type = "Task" [ 742.274022] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.287786] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230642, 'name': CreateSnapshot_Task, 'duration_secs': 0.527007} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.291605] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 742.291902] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230646, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.292676] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15377da0-1fdc-48c1-a9a2-69795862a167 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.345424] env[70013]: DEBUG nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 742.366498] env[70013]: DEBUG nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 742.382867] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 742.384676] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.384676] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 742.384676] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.384676] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 742.384676] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 742.385065] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 742.385065] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 742.385065] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 742.385160] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 742.385263] env[70013]: DEBUG nova.virt.hardware [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 742.389546] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce289534-b3cf-43ca-a08a-66f2de02b308 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.402700] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d475d4-e0ca-4d91-b8e0-1c3222b76892 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.436338] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573205} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.446233] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25/b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 742.446693] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 742.447442] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3fd0fa3-4258-4cb6-a92f-1274f6ac4045 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.457082] env[70013]: DEBUG oslo_vmware.api [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230645, 'name': PowerOnVM_Task, 'duration_secs': 0.497315} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.458838] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 742.459097] env[70013]: INFO nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Took 9.08 seconds to spawn the instance on the hypervisor. [ 742.459365] env[70013]: DEBUG nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 742.459614] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 742.459614] env[70013]: value = "task-4230647" [ 742.459614] env[70013]: _type = "Task" [ 742.459614] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.461166] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b860894f-0124-494b-95f9-de9f519da4a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.474062] env[70013]: DEBUG nova.network.neutron [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updated VIF entry in instance network info cache for port 0da6d475-040d-43e8-9139-88652285a5fe. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 742.474930] env[70013]: DEBUG nova.network.neutron [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updating instance_info_cache with network_info: [{"id": "0da6d475-040d-43e8-9139-88652285a5fe", "address": "fa:16:3e:1e:a7:37", "network": {"id": "1c5012af-830c-4001-9013-9e8cf32698bb", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-222945680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "719bc6e70189480dbef8b541a7a63d36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0da6d475-04", "ovs_interfaceid": "0da6d475-040d-43e8-9139-88652285a5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.477420] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.478272] env[70013]: DEBUG nova.scheduler.client.report [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 742.490868] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.794600] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230646, 'name': PowerOffVM_Task, 'duration_secs': 0.226843} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.794902] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 742.795098] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance '3b29c99a-0d55-40b1-a155-199b1ced2146' progress to 17 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 742.815582] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 742.818788] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-aced16cf-47db-4221-8030-5b29dc4d623a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.833612] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 742.833612] env[70013]: value = "task-4230648" [ 742.833612] env[70013]: _type = "Task" [ 742.833612] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.844069] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230648, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.854587] env[70013]: DEBUG nova.compute.manager [None req-587dc350-b54c-4eb3-9780-69efbdc5e6c6 tempest-ServerExternalEventsTest-584239149 tempest-ServerExternalEventsTest-584239149-project] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Received event network-changed {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 742.854806] env[70013]: DEBUG nova.compute.manager [None req-587dc350-b54c-4eb3-9780-69efbdc5e6c6 tempest-ServerExternalEventsTest-584239149 tempest-ServerExternalEventsTest-584239149-project] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Refreshing instance network info cache due to event network-changed. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 742.855040] env[70013]: DEBUG oslo_concurrency.lockutils [None req-587dc350-b54c-4eb3-9780-69efbdc5e6c6 tempest-ServerExternalEventsTest-584239149 tempest-ServerExternalEventsTest-584239149-project] Acquiring lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.855308] env[70013]: DEBUG oslo_concurrency.lockutils [None req-587dc350-b54c-4eb3-9780-69efbdc5e6c6 tempest-ServerExternalEventsTest-584239149 tempest-ServerExternalEventsTest-584239149-project] Acquired lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.857982] env[70013]: DEBUG nova.network.neutron [None req-587dc350-b54c-4eb3-9780-69efbdc5e6c6 tempest-ServerExternalEventsTest-584239149 tempest-ServerExternalEventsTest-584239149-project] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.913503] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.974725] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077456} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.975240] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.975947] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d679e6-13d4-4b16-9492-88dcbd903f23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.993584] env[70013]: DEBUG oslo_concurrency.lockutils [req-65cfc877-a514-41f2-bca8-7a38f482c635 req-08fc630f-abfd-4c4a-8444-8b84c81489dc service nova] Releasing lock "refresh_cache-8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.995589] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.685s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.995589] env[70013]: DEBUG nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 743.013300] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25/b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.015715] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.446s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.015945] env[70013]: DEBUG nova.objects.instance [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lazy-loading 'resources' on Instance uuid 417f8c74-81cc-47d6-80ca-b61d03ebec50 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 743.018056] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3142f486-d1a2-4a73-8862-7365340c7406 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.033989] env[70013]: INFO nova.compute.manager [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Took 25.39 seconds to build instance. [ 743.045982] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 743.045982] env[70013]: value = "task-4230649" [ 743.045982] env[70013]: _type = "Task" [ 743.045982] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.058587] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230649, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.228946] env[70013]: DEBUG nova.network.neutron [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Updated VIF entry in instance network info cache for port 47970e5c-2dab-40b7-abcc-f2bc90bcfad7. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 743.228946] env[70013]: DEBUG nova.network.neutron [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Updating instance_info_cache with network_info: [{"id": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "address": "fa:16:3e:03:9c:38", "network": {"id": "6055f3ca-7760-404c-8618-7f0e0cdc4237", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1361122985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eb984e7011845a1b0b848d8a17c6990", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47970e5c-2d", "ovs_interfaceid": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.309608] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 743.310479] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.310479] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 743.310479] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.310479] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 743.310838] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 743.310838] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 743.311772] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 743.311772] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 743.311772] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 743.311772] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 743.320458] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-169f2ce9-020e-4b70-83df-ab602db5b51c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.364330] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230648, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.365759] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 743.365759] env[70013]: value = "task-4230650" [ 743.365759] env[70013]: _type = "Task" [ 743.365759] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.380525] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230650, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.517465] env[70013]: DEBUG nova.compute.utils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 743.525257] env[70013]: DEBUG nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 743.525643] env[70013]: DEBUG nova.network.neutron [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 743.539039] env[70013]: DEBUG oslo_concurrency.lockutils [None req-35761083-c4b5-472b-9547-0a62c3397e71 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.298s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.574242] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230649, 'name': ReconfigVM_Task, 'duration_secs': 0.476207} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.574242] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Reconfigured VM instance instance-00000010 to attach disk [datastore1] b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25/b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.574242] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c819974-1126-490d-8d44-708dc2ca0772 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.584414] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 743.584414] env[70013]: value = "task-4230651" [ 743.584414] env[70013]: _type = "Task" [ 743.584414] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.592281] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6624fe-29a2-4d70-8b02-65d948c7d28b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.601589] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230651, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.606096] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab35e2e0-8561-489b-9e98-a72b59b2c63c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.643121] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616fd870-7f4a-4a88-b8ec-36ba975c8b0c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.652972] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5196af2a-549a-490c-9d6b-f0babab5a57c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.673016] env[70013]: DEBUG nova.compute.provider_tree [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.679150] env[70013]: DEBUG nova.policy [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e078550755304f82a0841fd93cff93b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c266552bee9a48dfa99379521fb8efdc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 743.730138] env[70013]: DEBUG oslo_concurrency.lockutils [req-70f5d1c7-97d5-40f1-993f-ef37ad20b0e8 req-78fcbb27-3fda-4d78-a3cc-abe042ba3980 service nova] Releasing lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.862019] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230648, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.875707] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230650, 'name': ReconfigVM_Task, 'duration_secs': 0.321013} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.879037] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance '3b29c99a-0d55-40b1-a155-199b1ced2146' progress to 33 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 743.882114] env[70013]: DEBUG nova.network.neutron [None req-587dc350-b54c-4eb3-9780-69efbdc5e6c6 tempest-ServerExternalEventsTest-584239149 tempest-ServerExternalEventsTest-584239149-project] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Updating instance_info_cache with network_info: [{"id": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "address": "fa:16:3e:b5:79:fd", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4a28327-a5", "ovs_interfaceid": "c4a28327-a59d-44df-9cc4-ee9bcd50c21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.915760] env[70013]: DEBUG nova.network.neutron [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Successfully updated port: 3e878689-ac09-469d-8490-a73b6e02d403 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 744.029570] env[70013]: DEBUG nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 744.046298] env[70013]: DEBUG nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 744.100103] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230651, 'name': Rename_Task, 'duration_secs': 0.206871} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.100456] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 744.100742] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77e855d2-8938-42c8-b76c-079d5a3a21fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.109477] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 744.109477] env[70013]: value = "task-4230656" [ 744.109477] env[70013]: _type = "Task" [ 744.109477] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.118402] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230656, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.179105] env[70013]: DEBUG nova.scheduler.client.report [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 744.363087] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230648, 'name': CloneVM_Task} progress is 95%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.387925] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 744.388351] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.388457] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 744.388743] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.388957] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 744.394279] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 744.394647] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 744.394916] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 744.395110] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 744.395303] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 744.395535] env[70013]: DEBUG nova.virt.hardware [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 744.403128] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Reconfiguring VM instance instance-00000006 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 744.403889] env[70013]: DEBUG oslo_concurrency.lockutils [None req-587dc350-b54c-4eb3-9780-69efbdc5e6c6 tempest-ServerExternalEventsTest-584239149 tempest-ServerExternalEventsTest-584239149-project] Releasing lock "refresh_cache-fc0b7f87-8a36-440e-b158-30b9d9920572" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.404380] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-805b4ede-1745-4fa1-b896-0a5b3936a910 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.426303] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "refresh_cache-f12b8575-5082-4be9-9bf5-f4279860d19d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.426303] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquired lock "refresh_cache-f12b8575-5082-4be9-9bf5-f4279860d19d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.426953] env[70013]: DEBUG nova.network.neutron [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.431840] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 744.431840] env[70013]: value = "task-4230657" [ 744.431840] env[70013]: _type = "Task" [ 744.431840] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.442861] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230657, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.584017] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.598801] env[70013]: DEBUG nova.network.neutron [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Successfully created port: 7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.627861] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230656, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.686302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.669s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.692019] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.886s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.692758] env[70013]: INFO nova.compute.claims [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.727884] env[70013]: INFO nova.scheduler.client.report [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Deleted allocations for instance 417f8c74-81cc-47d6-80ca-b61d03ebec50 [ 744.862896] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230648, 'name': CloneVM_Task, 'duration_secs': 1.844008} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.863987] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Created linked-clone VM from snapshot [ 744.865365] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f38a23-df04-46e1-8f84-6db1915c4955 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.881139] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Uploading image 584aacc9-664c-42e1-8636-52a852719f61 {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 744.917362] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 744.917362] env[70013]: value = "vm-837052" [ 744.917362] env[70013]: _type = "VirtualMachine" [ 744.917362] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 744.918720] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4a85606a-5f1d-4488-b1b6-29d6e35e9006 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.931599] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease: (returnval){ [ 744.931599] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dad884-dbe0-91a8-53f5-6c5f924bc4eb" [ 744.931599] env[70013]: _type = "HttpNfcLease" [ 744.931599] env[70013]: } obtained for exporting VM: (result){ [ 744.931599] env[70013]: value = "vm-837052" [ 744.931599] env[70013]: _type = "VirtualMachine" [ 744.931599] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 744.932042] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the lease: (returnval){ [ 744.932042] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dad884-dbe0-91a8-53f5-6c5f924bc4eb" [ 744.932042] env[70013]: _type = "HttpNfcLease" [ 744.932042] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 744.948028] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 744.948028] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dad884-dbe0-91a8-53f5-6c5f924bc4eb" [ 744.948028] env[70013]: _type = "HttpNfcLease" [ 744.948028] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 744.950665] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230657, 'name': ReconfigVM_Task, 'duration_secs': 0.225627} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.951043] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Reconfigured VM instance instance-00000006 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 744.952039] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113f9a3d-9ac6-4ebd-84e7-14e35a3bdea5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.985903] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] 3b29c99a-0d55-40b1-a155-199b1ced2146/3b29c99a-0d55-40b1-a155-199b1ced2146.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.986895] env[70013]: DEBUG nova.network.neutron [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.989442] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a32674d3-7acc-493c-a271-839b831188d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.018981] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 745.018981] env[70013]: value = "task-4230659" [ 745.018981] env[70013]: _type = "Task" [ 745.018981] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.026148] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230659, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.044608] env[70013]: DEBUG nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 745.092582] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:33:26Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2122098332',id=28,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1154722654',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 745.092582] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.092582] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 745.092840] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.092840] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 745.092840] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 745.092840] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 745.092840] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 745.093076] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 745.093076] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 745.093142] env[70013]: DEBUG nova.virt.hardware [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 745.094616] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b68c74b-fdb6-40ac-ad42-da25708c5be2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.106257] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae7b58b-4de3-4278-8059-d4f6fe95c05b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.133609] env[70013]: DEBUG oslo_vmware.api [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230656, 'name': PowerOnVM_Task, 'duration_secs': 0.622436} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.133609] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.133609] env[70013]: INFO nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Took 9.04 seconds to spawn the instance on the hypervisor. [ 745.133837] env[70013]: DEBUG nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 745.134752] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08b0cc7-9405-4301-9e5e-369d79b62968 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.239672] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fa7e70c-f994-4864-8036-1ea70c1ff9f2 tempest-ServerAddressesNegativeTestJSON-943476673 tempest-ServerAddressesNegativeTestJSON-943476673-project-member] Lock "417f8c74-81cc-47d6-80ca-b61d03ebec50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.917s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 745.347897] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "fc0b7f87-8a36-440e-b158-30b9d9920572" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.348416] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.348626] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "fc0b7f87-8a36-440e-b158-30b9d9920572-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.348932] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.349045] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 745.352229] env[70013]: INFO nova.compute.manager [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Terminating instance [ 745.419500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Acquiring lock "8317b86e-1408-484c-822a-3b2f248c3a4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.419500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Lock "8317b86e-1408-484c-822a-3b2f248c3a4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.448440] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 745.448440] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dad884-dbe0-91a8-53f5-6c5f924bc4eb" [ 745.448440] env[70013]: _type = "HttpNfcLease" [ 745.448440] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 745.448440] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 745.448440] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dad884-dbe0-91a8-53f5-6c5f924bc4eb" [ 745.448440] env[70013]: _type = "HttpNfcLease" [ 745.448440] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 745.449225] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25598c5c-f2f8-4088-96aa-ba6aca6c5ff8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.460511] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5202e146-d6c6-5b72-f5e4-11ccae904463/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 745.460774] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5202e146-d6c6-5b72-f5e4-11ccae904463/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 745.538021] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230659, 'name': ReconfigVM_Task, 'duration_secs': 0.407021} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.538021] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Reconfigured VM instance instance-00000006 to attach disk [datastore1] 3b29c99a-0d55-40b1-a155-199b1ced2146/3b29c99a-0d55-40b1-a155-199b1ced2146.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.538021] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance '3b29c99a-0d55-40b1-a155-199b1ced2146' progress to 50 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 745.556479] env[70013]: DEBUG nova.compute.manager [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Received event network-vif-plugged-3e878689-ac09-469d-8490-a73b6e02d403 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 745.556894] env[70013]: DEBUG oslo_concurrency.lockutils [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] Acquiring lock "f12b8575-5082-4be9-9bf5-f4279860d19d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.557180] env[70013]: DEBUG oslo_concurrency.lockutils [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.559826] env[70013]: DEBUG oslo_concurrency.lockutils [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 745.559826] env[70013]: DEBUG nova.compute.manager [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] No waiting events found dispatching network-vif-plugged-3e878689-ac09-469d-8490-a73b6e02d403 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 745.559826] env[70013]: WARNING nova.compute.manager [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Received unexpected event network-vif-plugged-3e878689-ac09-469d-8490-a73b6e02d403 for instance with vm_state building and task_state spawning. [ 745.559826] env[70013]: DEBUG nova.compute.manager [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Received event network-changed-3e878689-ac09-469d-8490-a73b6e02d403 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 745.559826] env[70013]: DEBUG nova.compute.manager [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Refreshing instance network info cache due to event network-changed-3e878689-ac09-469d-8490-a73b6e02d403. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 745.560446] env[70013]: DEBUG oslo_concurrency.lockutils [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] Acquiring lock "refresh_cache-f12b8575-5082-4be9-9bf5-f4279860d19d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.588788] env[70013]: DEBUG nova.network.neutron [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Updating instance_info_cache with network_info: [{"id": "3e878689-ac09-469d-8490-a73b6e02d403", "address": "fa:16:3e:b8:3e:81", "network": {"id": "891185a3-4f05-4fbf-812b-16a46ee30b3b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-148731314-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0dd47358d7b5432d93bddf797588a0cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba6157eb-73cb-428a-9f46-99081165d7eb", "external-id": "nsx-vlan-transportzone-463", "segmentation_id": 463, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e878689-ac", "ovs_interfaceid": "3e878689-ac09-469d-8490-a73b6e02d403", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.667362] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-358b2669-ca0f-408a-a271-d273266e91ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.669319] env[70013]: INFO nova.compute.manager [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Took 27.89 seconds to build instance. [ 745.681409] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "fc473d40-b57b-437e-9511-58a0ba700a69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.681409] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.681409] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.681638] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.681638] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 745.685704] env[70013]: INFO nova.compute.manager [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Terminating instance [ 745.857444] env[70013]: DEBUG nova.compute.manager [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 745.857637] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 745.860015] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45e6472-ed14-42be-a91a-eeadf014d4fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.870690] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 745.871260] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d716448b-1052-4106-b399-7efae2981340 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.882602] env[70013]: DEBUG oslo_vmware.api [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 745.882602] env[70013]: value = "task-4230660" [ 745.882602] env[70013]: _type = "Task" [ 745.882602] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.894968] env[70013]: DEBUG oslo_vmware.api [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.049582] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da56bfc-5c19-4a00-842b-ca919487d942 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.082444] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70bb8922-fcb3-495a-8a88-3ba0a37d1778 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.134256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Releasing lock "refresh_cache-f12b8575-5082-4be9-9bf5-f4279860d19d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 746.134256] env[70013]: DEBUG nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Instance network_info: |[{"id": "3e878689-ac09-469d-8490-a73b6e02d403", "address": "fa:16:3e:b8:3e:81", "network": {"id": "891185a3-4f05-4fbf-812b-16a46ee30b3b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-148731314-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0dd47358d7b5432d93bddf797588a0cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba6157eb-73cb-428a-9f46-99081165d7eb", "external-id": "nsx-vlan-transportzone-463", "segmentation_id": 463, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e878689-ac", "ovs_interfaceid": "3e878689-ac09-469d-8490-a73b6e02d403", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 746.134423] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance '3b29c99a-0d55-40b1-a155-199b1ced2146' progress to 67 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 746.147178] env[70013]: DEBUG oslo_concurrency.lockutils [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] Acquired lock "refresh_cache-f12b8575-5082-4be9-9bf5-f4279860d19d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 746.147858] env[70013]: DEBUG nova.network.neutron [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Refreshing network info cache for port 3e878689-ac09-469d-8490-a73b6e02d403 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 746.151254] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:3e:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ba6157eb-73cb-428a-9f46-99081165d7eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e878689-ac09-469d-8490-a73b6e02d403', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.173014] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Creating folder: Project (0dd47358d7b5432d93bddf797588a0cb). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.182373] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5beb0c8-d5f9-4574-810c-53d3fbadada3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.187908] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ad06ef3-4878-4fd4-8723-875c2bb2817d tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.022s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.195544] env[70013]: DEBUG nova.compute.manager [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 746.195544] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 746.195934] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106380ae-13e5-4fe5-83b0-50778be3b6ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.208457] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Created folder: Project (0dd47358d7b5432d93bddf797588a0cb) in parent group-v836999. [ 746.208863] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Creating folder: Instances. Parent ref: group-v837056. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.216066] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a19a59fe-a270-405a-babb-c9dc7fcfda40 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.219165] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 746.221351] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88a2670f-41b6-4843-9ac2-b0e381ad3db7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.231077] env[70013]: DEBUG oslo_vmware.api [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 746.231077] env[70013]: value = "task-4230663" [ 746.231077] env[70013]: _type = "Task" [ 746.231077] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.239656] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Created folder: Instances in parent group-v837056. [ 746.239656] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 746.240766] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 746.244793] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5eb5da1-bf8b-419d-9a8f-5e8737e953fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.270413] env[70013]: DEBUG oslo_vmware.api [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.277460] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.277460] env[70013]: value = "task-4230664" [ 746.277460] env[70013]: _type = "Task" [ 746.277460] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.291101] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230664, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.334734] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a4cc1c-530b-4094-89ce-c617444f5040 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.345625] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc2a9d2-67f0-4fe1-a1e8-2897b92b6a77 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.391295] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cafd09-4535-4f39-b150-9bd2a2fffaac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.405773] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2518171f-8c4c-4560-be25-eabba22ff8f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.411163] env[70013]: DEBUG oslo_vmware.api [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230660, 'name': PowerOffVM_Task, 'duration_secs': 0.283602} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.411806] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 746.411806] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 746.412906] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-04553ac6-c98e-4abf-9d67-60cd87f40078 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.424495] env[70013]: DEBUG nova.compute.provider_tree [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.513685] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 746.513685] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 746.513685] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Deleting the datastore file [datastore1] fc0b7f87-8a36-440e-b158-30b9d9920572 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 746.513915] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c2e1fc7-c27d-48a5-97e8-7f3d272a1acf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.521902] env[70013]: DEBUG oslo_vmware.api [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for the task: (returnval){ [ 746.521902] env[70013]: value = "task-4230666" [ 746.521902] env[70013]: _type = "Task" [ 746.521902] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.535928] env[70013]: DEBUG oslo_vmware.api [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230666, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.693193] env[70013]: DEBUG nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 746.748429] env[70013]: DEBUG oslo_vmware.api [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230663, 'name': PowerOffVM_Task, 'duration_secs': 0.274853} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.748792] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 746.748958] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 746.749243] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-274306f9-4fe2-4784-ba7d-1764bb8e3420 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.797119] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230664, 'name': CreateVM_Task, 'duration_secs': 0.452098} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.797119] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 746.797119] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.797119] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 746.797119] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 746.797831] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06c37c6f-d77b-443b-85ce-ac3f7ed6b1c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.805259] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 746.805259] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52865e13-0b25-f27c-d13d-0e271bff2801" [ 746.805259] env[70013]: _type = "Task" [ 746.805259] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.815629] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52865e13-0b25-f27c-d13d-0e271bff2801, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.828066] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 746.828329] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 746.829056] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleting the datastore file [datastore1] fc473d40-b57b-437e-9511-58a0ba700a69 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 746.829056] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-765b789f-43b4-4807-9c43-071b581e3751 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.837451] env[70013]: DEBUG oslo_vmware.api [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 746.837451] env[70013]: value = "task-4230668" [ 746.837451] env[70013]: _type = "Task" [ 746.837451] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.849512] env[70013]: DEBUG oslo_vmware.api [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230668, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.890883] env[70013]: DEBUG nova.network.neutron [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Port b402fce4-7392-4139-8de5-7620820e0814 binding to destination host cpu-1 is already ACTIVE {{(pid=70013) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 746.929667] env[70013]: DEBUG nova.scheduler.client.report [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 747.047201] env[70013]: DEBUG oslo_vmware.api [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Task: {'id': task-4230666, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210681} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.047494] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 747.047739] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 747.048000] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 747.048655] env[70013]: INFO nova.compute.manager [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Took 1.19 seconds to destroy the instance on the hypervisor. [ 747.049171] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 747.049171] env[70013]: DEBUG nova.compute.manager [-] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 747.049267] env[70013]: DEBUG nova.network.neutron [-] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.236811] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.238437] env[70013]: DEBUG nova.network.neutron [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Updated VIF entry in instance network info cache for port 3e878689-ac09-469d-8490-a73b6e02d403. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 747.238437] env[70013]: DEBUG nova.network.neutron [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Updating instance_info_cache with network_info: [{"id": "3e878689-ac09-469d-8490-a73b6e02d403", "address": "fa:16:3e:b8:3e:81", "network": {"id": "891185a3-4f05-4fbf-812b-16a46ee30b3b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-148731314-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0dd47358d7b5432d93bddf797588a0cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba6157eb-73cb-428a-9f46-99081165d7eb", "external-id": "nsx-vlan-transportzone-463", "segmentation_id": 463, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e878689-ac", "ovs_interfaceid": "3e878689-ac09-469d-8490-a73b6e02d403", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.322265] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52865e13-0b25-f27c-d13d-0e271bff2801, 'name': SearchDatastore_Task, 'duration_secs': 0.02221} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.325330] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.325971] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 747.326413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.326633] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.326832] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.327553] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70047329-db8e-4e48-8dc1-b333408d656d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.347533] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.347876] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 747.349970] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d56e513-943d-4a80-bcac-b2b0482e7630 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.359779] env[70013]: DEBUG oslo_vmware.api [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4230668, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176115} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.361941] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 747.362687] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 747.363585] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 747.363585] env[70013]: INFO nova.compute.manager [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Took 1.17 seconds to destroy the instance on the hypervisor. [ 747.364256] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 747.365688] env[70013]: DEBUG nova.compute.manager [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 747.365941] env[70013]: DEBUG nova.network.neutron [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.368148] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 747.368148] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528d61f2-3aaf-6882-187c-3dcc9f6aca5a" [ 747.368148] env[70013]: _type = "Task" [ 747.368148] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.379737] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528d61f2-3aaf-6882-187c-3dcc9f6aca5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.436039] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.745s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.438022] env[70013]: DEBUG nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 747.444018] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.017s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.445946] env[70013]: INFO nova.compute.claims [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.742426] env[70013]: DEBUG oslo_concurrency.lockutils [req-d8d189f6-5ac6-4982-9d4d-2220cc573eaf req-c32df33a-b40e-4aff-9801-8c0291a304c2 service nova] Releasing lock "refresh_cache-f12b8575-5082-4be9-9bf5-f4279860d19d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.881382] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528d61f2-3aaf-6882-187c-3dcc9f6aca5a, 'name': SearchDatastore_Task, 'duration_secs': 0.040516} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.882500] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-951229dc-8b19-4970-a386-4865d1091aa5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.894378] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 747.894378] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521d97ad-3bf6-e6bf-119b-fa784cfe5a7d" [ 747.894378] env[70013]: _type = "Task" [ 747.894378] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.916794] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521d97ad-3bf6-e6bf-119b-fa784cfe5a7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.928605] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.928605] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.928605] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.954262] env[70013]: DEBUG nova.compute.utils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 747.964254] env[70013]: DEBUG nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 747.964985] env[70013]: DEBUG nova.network.neutron [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.382271] env[70013]: DEBUG nova.policy [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1dba1615f64547ae9700e5264743248e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '64c4d90bd33e4c82be86c8c33a1f6826', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 748.413729] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521d97ad-3bf6-e6bf-119b-fa784cfe5a7d, 'name': SearchDatastore_Task, 'duration_secs': 0.010567} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.414014] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.414291] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f12b8575-5082-4be9-9bf5-f4279860d19d/f12b8575-5082-4be9-9bf5-f4279860d19d.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 748.414563] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac88cc36-c755-4caa-8484-ca77cc7fcd80 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.424106] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 748.424106] env[70013]: value = "task-4230670" [ 748.424106] env[70013]: _type = "Task" [ 748.424106] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.437917] env[70013]: DEBUG nova.network.neutron [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Successfully updated port: 7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 748.460682] env[70013]: DEBUG nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 748.470281] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.939272] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230670, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.956859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.957354] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquired lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 748.958446] env[70013]: DEBUG nova.network.neutron [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.959635] env[70013]: DEBUG nova.network.neutron [-] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.075305] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494ff485-f140-4267-b26b-c34672ddb70e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.088013] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3a1d9b-a1f8-4c09-ac6d-33a021492b64 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.129665] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43552d11-690e-444b-af38-f2fdd9d3c743 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.140117] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d21bb8a-b741-4557-9655-2e2cef4c91eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.153171] env[70013]: DEBUG nova.compute.provider_tree [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.291395] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.291576] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.291755] env[70013]: DEBUG nova.network.neutron [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.329721] env[70013]: DEBUG nova.network.neutron [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.442016] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230670, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.468552] env[70013]: INFO nova.compute.manager [-] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Took 2.42 seconds to deallocate network for instance. [ 749.480892] env[70013]: DEBUG nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 749.511327] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 749.513505] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.513505] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 749.513505] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.513505] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 749.513505] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 749.516165] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 749.516165] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 749.516165] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 749.516165] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 749.516165] env[70013]: DEBUG nova.virt.hardware [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 749.516645] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8367b8b8-aba0-4ed2-8274-feb57d4dd13f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.523234] env[70013]: DEBUG nova.compute.manager [req-aeabc726-0293-4ac1-b8a7-ae39bf108ed2 req-9707bde7-4ffe-42a5-aa93-8d629278725b service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Received event network-vif-plugged-7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 749.523332] env[70013]: DEBUG oslo_concurrency.lockutils [req-aeabc726-0293-4ac1-b8a7-ae39bf108ed2 req-9707bde7-4ffe-42a5-aa93-8d629278725b service nova] Acquiring lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.523602] env[70013]: DEBUG oslo_concurrency.lockutils [req-aeabc726-0293-4ac1-b8a7-ae39bf108ed2 req-9707bde7-4ffe-42a5-aa93-8d629278725b service nova] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 749.523759] env[70013]: DEBUG oslo_concurrency.lockutils [req-aeabc726-0293-4ac1-b8a7-ae39bf108ed2 req-9707bde7-4ffe-42a5-aa93-8d629278725b service nova] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.523968] env[70013]: DEBUG nova.compute.manager [req-aeabc726-0293-4ac1-b8a7-ae39bf108ed2 req-9707bde7-4ffe-42a5-aa93-8d629278725b service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] No waiting events found dispatching network-vif-plugged-7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 749.524371] env[70013]: WARNING nova.compute.manager [req-aeabc726-0293-4ac1-b8a7-ae39bf108ed2 req-9707bde7-4ffe-42a5-aa93-8d629278725b service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Received unexpected event network-vif-plugged-7634504a-a16c-44ea-9a82-e8f88bcc65ea for instance with vm_state building and task_state spawning. [ 749.533588] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afe9683-2239-4820-bcdf-5914cd473d6b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.574302] env[70013]: DEBUG nova.network.neutron [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.657887] env[70013]: DEBUG nova.scheduler.client.report [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 749.832993] env[70013]: INFO nova.compute.manager [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Took 2.47 seconds to deallocate network for instance. [ 749.900456] env[70013]: DEBUG nova.compute.manager [req-cbb198de-cf3a-4a2c-8981-82c65df111ef req-640978e3-84bd-4b6f-bd4c-cd2af23b7e6e service nova] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Received event network-vif-deleted-c4a28327-a59d-44df-9cc4-ee9bcd50c21e {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 749.939105] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230670, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.985869] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.169172] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.724s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.169172] env[70013]: DEBUG nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 750.172419] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.596s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.172696] env[70013]: DEBUG nova.objects.instance [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lazy-loading 'resources' on Instance uuid 9a553d85-5086-46c5-8df4-451928e38ed9 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.345510] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.377067] env[70013]: DEBUG nova.network.neutron [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Updating instance_info_cache with network_info: [{"id": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "address": "fa:16:3e:79:3b:a7", "network": {"id": "dabcb531-f742-4f77-a8b8-97269090c723", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1602634587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c266552bee9a48dfa99379521fb8efdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7634504a-a1", "ovs_interfaceid": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.444026] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230670, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.730813} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.446244] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f12b8575-5082-4be9-9bf5-f4279860d19d/f12b8575-5082-4be9-9bf5-f4279860d19d.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 750.446244] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.446244] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e0c9b58-f7e9-4222-88d6-6879e3043fc9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.457200] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 750.457200] env[70013]: value = "task-4230672" [ 750.457200] env[70013]: _type = "Task" [ 750.457200] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.475323] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230672, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.680954] env[70013]: DEBUG nova.compute.utils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 750.685763] env[70013]: DEBUG nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 750.685763] env[70013]: DEBUG nova.network.neutron [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 750.721752] env[70013]: DEBUG nova.network.neutron [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Successfully created port: 5708c0e5-6147-4b18-b294-885738a29578 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.741024] env[70013]: DEBUG nova.network.neutron [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance_info_cache with network_info: [{"id": "b402fce4-7392-4139-8de5-7620820e0814", "address": "fa:16:3e:44:ef:b6", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb402fce4-73", "ovs_interfaceid": "b402fce4-7392-4139-8de5-7620820e0814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.856353] env[70013]: DEBUG nova.policy [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 750.881879] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Releasing lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.882263] env[70013]: DEBUG nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Instance network_info: |[{"id": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "address": "fa:16:3e:79:3b:a7", "network": {"id": "dabcb531-f742-4f77-a8b8-97269090c723", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1602634587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c266552bee9a48dfa99379521fb8efdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7634504a-a1", "ovs_interfaceid": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 750.883444] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:3b:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b83383f-ed7a-4efd-aef7-aa8c15649d07', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7634504a-a16c-44ea-9a82-e8f88bcc65ea', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 750.903497] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Creating folder: Project (c266552bee9a48dfa99379521fb8efdc). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 750.905178] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14ec49dd-6f21-4221-acfc-2351c46a7508 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.920360] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Created folder: Project (c266552bee9a48dfa99379521fb8efdc) in parent group-v836999. [ 750.920360] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Creating folder: Instances. Parent ref: group-v837059. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 750.920360] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-368fdf16-ea2f-4f02-bc81-8bf1089f56b7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.935584] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Created folder: Instances in parent group-v837059. [ 750.936610] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 750.936610] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 750.936610] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91abeba2-eb8b-465e-b285-0e7d068ec218 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.975734] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 750.975734] env[70013]: value = "task-4230676" [ 750.975734] env[70013]: _type = "Task" [ 750.975734] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.976195] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230672, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077439} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.976403] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 750.980823] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f01355-698a-4a30-99a4-3dbc5c7adbd7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.009417] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230676, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.023155] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] f12b8575-5082-4be9-9bf5-f4279860d19d/f12b8575-5082-4be9-9bf5-f4279860d19d.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.028334] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65cd1293-001c-4d68-af08-4ce1158c0f55 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.053544] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 751.053544] env[70013]: value = "task-4230677" [ 751.053544] env[70013]: _type = "Task" [ 751.053544] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.066617] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230677, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.191957] env[70013]: DEBUG nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 751.245363] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.279026] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74cb15a5-ce57-4f63-a019-39809a06a41a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.293199] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3ad59c-fb8f-42b6-a747-675e553ea7c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.335284] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6dbf5b-fc75-4a47-b0e4-7c3439ef0f28 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.349862] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351f3101-f876-4b79-a8aa-3153b290b7de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.383037] env[70013]: DEBUG nova.compute.provider_tree [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.492720] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230676, 'name': CreateVM_Task, 'duration_secs': 0.405655} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.492966] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 751.494143] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.494143] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.494287] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 751.495081] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb5c3009-a59f-48be-9faf-c901b969fd2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.502079] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 751.502079] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5268d8c3-539c-cb1e-00e5-d89bb7ae0f77" [ 751.502079] env[70013]: _type = "Task" [ 751.502079] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.513062] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5268d8c3-539c-cb1e-00e5-d89bb7ae0f77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.564923] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230677, 'name': ReconfigVM_Task, 'duration_secs': 0.437819} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.565306] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Reconfigured VM instance instance-00000011 to attach disk [datastore2] f12b8575-5082-4be9-9bf5-f4279860d19d/f12b8575-5082-4be9-9bf5-f4279860d19d.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 751.566202] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93a6a2f6-8918-4e2b-87ea-8c8b4c696937 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.574557] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 751.574557] env[70013]: value = "task-4230678" [ 751.574557] env[70013]: _type = "Task" [ 751.574557] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.583661] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230678, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.803798] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1ddedd-eaef-46be-8463-c349cd5bd8c3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.815490] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Acquiring lock "b0aad98e-d56c-403c-b075-ddd1634516fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.815740] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Lock "b0aad98e-d56c-403c-b075-ddd1634516fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.835705] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722c6b41-e0a9-4cd3-98de-1f4210554f23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.848130] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance '3b29c99a-0d55-40b1-a155-199b1ced2146' progress to 83 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 751.886779] env[70013]: DEBUG nova.scheduler.client.report [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 752.015355] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5268d8c3-539c-cb1e-00e5-d89bb7ae0f77, 'name': SearchDatastore_Task, 'duration_secs': 0.011585} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.015786] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.016016] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.016301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.016416] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 752.016599] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.016903] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20e3b59d-c947-4824-87bd-f209729ebc74 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.027671] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.027885] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 752.028750] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07be1ab7-07b2-45ca-a162-4203628d7bab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.035515] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 752.035515] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1b6d0-31c7-ba3b-8ee5-475b22704872" [ 752.035515] env[70013]: _type = "Task" [ 752.035515] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.046996] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1b6d0-31c7-ba3b-8ee5-475b22704872, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.091237] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230678, 'name': Rename_Task, 'duration_secs': 0.214265} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.091619] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.091999] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32a835b7-989e-49cf-9af8-9220c1a13ee5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.100690] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 752.100690] env[70013]: value = "task-4230679" [ 752.100690] env[70013]: _type = "Task" [ 752.100690] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.111287] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.203713] env[70013]: DEBUG nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 752.255590] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 752.255948] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.256164] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 752.256398] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.256491] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 752.256675] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 752.256946] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 752.257146] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 752.257359] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 752.257557] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 752.257746] env[70013]: DEBUG nova.virt.hardware [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 752.258903] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7f24f9-20f4-471d-8647-a53ee6038cc9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.272447] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f68144-afd8-47a2-8dee-70ed9abcb09e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.311287] env[70013]: DEBUG nova.network.neutron [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Successfully created port: 3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.357241] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.358070] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a87dc95-4c53-4cb2-acf6-0de9673a7d22 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.366451] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 752.366451] env[70013]: value = "task-4230680" [ 752.366451] env[70013]: _type = "Task" [ 752.366451] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.379945] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.395511] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.223s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.398766] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.971s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.400904] env[70013]: INFO nova.compute.claims [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.442887] env[70013]: INFO nova.scheduler.client.report [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Deleted allocations for instance 9a553d85-5086-46c5-8df4-451928e38ed9 [ 752.549875] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1b6d0-31c7-ba3b-8ee5-475b22704872, 'name': SearchDatastore_Task, 'duration_secs': 0.009803} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.550820] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0da6765f-5fe7-4ada-a50d-4b35a1bda72f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.559911] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 752.559911] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52593218-95ff-714b-434d-1abc6bf3342b" [ 752.559911] env[70013]: _type = "Task" [ 752.559911] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.572685] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52593218-95ff-714b-434d-1abc6bf3342b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.615982] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230679, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.884749] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230680, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.956837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aad0a085-a00d-4c57-980d-fc4fa9140110 tempest-InstanceActionsV221TestJSON-840280228 tempest-InstanceActionsV221TestJSON-840280228-project-member] Lock "9a553d85-5086-46c5-8df4-451928e38ed9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.865s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.080789] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52593218-95ff-714b-434d-1abc6bf3342b, 'name': SearchDatastore_Task, 'duration_secs': 0.013443} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.083868] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.083868] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] b23b4e67-67a6-4117-9a62-3c0dcd25e828/b23b4e67-67a6-4117-9a62-3c0dcd25e828.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 753.083868] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2477ec13-38d3-4cb1-ac11-5c8ec51d8880 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.092964] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 753.092964] env[70013]: value = "task-4230681" [ 753.092964] env[70013]: _type = "Task" [ 753.092964] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.110192] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230681, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.124635] env[70013]: DEBUG oslo_vmware.api [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230679, 'name': PowerOnVM_Task, 'duration_secs': 0.623154} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.125634] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.127092] env[70013]: INFO nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Took 10.78 seconds to spawn the instance on the hypervisor. [ 753.127092] env[70013]: DEBUG nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 753.127497] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7be68c1-e605-42ed-9722-4e309d84da38 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.387023] env[70013]: DEBUG oslo_vmware.api [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230680, 'name': PowerOnVM_Task, 'duration_secs': 0.935577} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.387416] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.387620] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4e74b76b-b6f1-4e3d-a008-319e0ef60aa6 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance '3b29c99a-0d55-40b1-a155-199b1ced2146' progress to 100 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 753.486585] env[70013]: DEBUG nova.compute.manager [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Received event network-changed-7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 753.488670] env[70013]: DEBUG nova.compute.manager [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Refreshing instance network info cache due to event network-changed-7634504a-a16c-44ea-9a82-e8f88bcc65ea. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 753.493027] env[70013]: DEBUG oslo_concurrency.lockutils [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] Acquiring lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.493027] env[70013]: DEBUG oslo_concurrency.lockutils [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] Acquired lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.493027] env[70013]: DEBUG nova.network.neutron [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Refreshing network info cache for port 7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.592216] env[70013]: DEBUG nova.compute.manager [req-ef038a0c-0d93-49f4-afe1-6ed5c210bcda req-6ffde0de-0ef9-4ec5-a441-23c8e9a51c6e service nova] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Received event network-vif-deleted-4f1ea900-8772-4876-82d4-0c75f8ab8e22 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 753.613899] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230681, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.664591] env[70013]: INFO nova.compute.manager [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Took 32.46 seconds to build instance. [ 753.994345] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246ed26e-c782-4837-9930-7daa5e5dd31b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.008376] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375a3538-57d7-417f-95f8-6499e2d16921 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.048274] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fd2172-ac5d-4e7b-8956-06c4f1dce544 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.058473] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5202e146-d6c6-5b72-f5e4-11ccae904463/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 754.060158] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf4be7a-dae9-461c-8a76-bf1d74824f3d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.065992] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8602b445-01d5-4329-be4d-9553723be161 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.083780] env[70013]: DEBUG nova.compute.provider_tree [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.086855] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5202e146-d6c6-5b72-f5e4-11ccae904463/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 754.087109] env[70013]: ERROR oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5202e146-d6c6-5b72-f5e4-11ccae904463/disk-0.vmdk due to incomplete transfer. [ 754.087628] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1a20d0bd-05e7-447e-b310-ccbc01ab1646 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.101584] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5202e146-d6c6-5b72-f5e4-11ccae904463/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 754.101584] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Uploaded image 584aacc9-664c-42e1-8636-52a852719f61 to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 754.103040] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 754.104239] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1cbfd8a6-eecd-40dd-8e54-c9d195f84684 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.111525] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230681, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.118217] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 754.118217] env[70013]: value = "task-4230683" [ 754.118217] env[70013]: _type = "Task" [ 754.118217] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.128203] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230683, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.168008] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bca2ff74-ed5a-4954-8611-0f71131688a7 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.067s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 754.365130] env[70013]: DEBUG nova.network.neutron [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Successfully updated port: 5708c0e5-6147-4b18-b294-885738a29578 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 754.589699] env[70013]: DEBUG nova.scheduler.client.report [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 754.609359] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230681, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.628050] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230683, 'name': Destroy_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.669848] env[70013]: DEBUG nova.network.neutron [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Updated VIF entry in instance network info cache for port 7634504a-a16c-44ea-9a82-e8f88bcc65ea. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 754.670991] env[70013]: DEBUG nova.network.neutron [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Updating instance_info_cache with network_info: [{"id": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "address": "fa:16:3e:79:3b:a7", "network": {"id": "dabcb531-f742-4f77-a8b8-97269090c723", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1602634587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c266552bee9a48dfa99379521fb8efdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7634504a-a1", "ovs_interfaceid": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.672508] env[70013]: DEBUG nova.compute.manager [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 754.866091] env[70013]: DEBUG nova.network.neutron [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Successfully updated port: 3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 754.870743] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "refresh_cache-3e938bb3-54a6-49f8-809d-d5aee0349eae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.870743] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquired lock "refresh_cache-3e938bb3-54a6-49f8-809d-d5aee0349eae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.870743] env[70013]: DEBUG nova.network.neutron [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.096221] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.697s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.097316] env[70013]: DEBUG nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 755.106023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.696s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.106023] env[70013]: DEBUG nova.objects.instance [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 755.123097] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230681, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.57831} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.130738] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] b23b4e67-67a6-4117-9a62-3c0dcd25e828/b23b4e67-67a6-4117-9a62-3c0dcd25e828.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 755.130992] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.131364] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ec5b2c9-107c-4034-9d64-0e5302eb3ffe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.140646] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230683, 'name': Destroy_Task, 'duration_secs': 0.68531} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.145316] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Destroyed the VM [ 755.145316] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 755.145316] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 755.145316] env[70013]: value = "task-4230684" [ 755.145316] env[70013]: _type = "Task" [ 755.145316] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.145316] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cfd1a7f3-dbfa-4882-b895-9ee7004dccdb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.161025] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230684, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.161730] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 755.161730] env[70013]: value = "task-4230685" [ 755.161730] env[70013]: _type = "Task" [ 755.161730] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.175625] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230685, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.176573] env[70013]: DEBUG oslo_concurrency.lockutils [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] Releasing lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.176931] env[70013]: DEBUG nova.compute.manager [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Received event network-changed-47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 755.177234] env[70013]: DEBUG nova.compute.manager [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Refreshing instance network info cache due to event network-changed-47970e5c-2dab-40b7-abcc-f2bc90bcfad7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 755.177660] env[70013]: DEBUG oslo_concurrency.lockutils [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] Acquiring lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.177774] env[70013]: DEBUG oslo_concurrency.lockutils [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] Acquired lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.178930] env[70013]: DEBUG nova.network.neutron [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Refreshing network info cache for port 47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.208616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.373356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.373506] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.373705] env[70013]: DEBUG nova.network.neutron [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.432166] env[70013]: DEBUG nova.network.neutron [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.620012] env[70013]: DEBUG nova.compute.utils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 755.624031] env[70013]: DEBUG nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 755.664451] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230684, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072999} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.669085] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.671013] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b30162b-45c5-4fbd-bf96-83e551f0d5ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.702738] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] b23b4e67-67a6-4117-9a62-3c0dcd25e828/b23b4e67-67a6-4117-9a62-3c0dcd25e828.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.707530] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f61da316-9d0a-4025-bd85-0d5d81e65875 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.723080] env[70013]: DEBUG oslo_vmware.api [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230685, 'name': RemoveSnapshot_Task, 'duration_secs': 0.474258} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.724381] env[70013]: DEBUG nova.network.neutron [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Updating instance_info_cache with network_info: [{"id": "5708c0e5-6147-4b18-b294-885738a29578", "address": "fa:16:3e:5b:99:3b", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5708c0e5-61", "ovs_interfaceid": "5708c0e5-6147-4b18-b294-885738a29578", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.726043] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 755.726302] env[70013]: INFO nova.compute.manager [None req-5c28eccb-4429-469e-909f-612ffbc568e8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Took 15.01 seconds to snapshot the instance on the hypervisor. [ 755.736038] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 755.736038] env[70013]: value = "task-4230687" [ 755.736038] env[70013]: _type = "Task" [ 755.736038] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.751253] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.124016] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07a31b3e-a721-46a7-a8d1-8b7b79c6c97a tempest-ServersAdmin275Test-393927465 tempest-ServersAdmin275Test-393927465-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.125290] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.409s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.127869] env[70013]: INFO nova.compute.claims [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.131485] env[70013]: DEBUG nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 756.161743] env[70013]: DEBUG nova.network.neutron [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.230659] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Releasing lock "refresh_cache-3e938bb3-54a6-49f8-809d-d5aee0349eae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.231047] env[70013]: DEBUG nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Instance network_info: |[{"id": "5708c0e5-6147-4b18-b294-885738a29578", "address": "fa:16:3e:5b:99:3b", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5708c0e5-61", "ovs_interfaceid": "5708c0e5-6147-4b18-b294-885738a29578", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 756.231693] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:99:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5708c0e5-6147-4b18-b294-885738a29578', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 756.241892] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Creating folder: Project (64c4d90bd33e4c82be86c8c33a1f6826). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 756.244802] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97283325-3eef-44f0-9e04-6fd96d23670f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.259281] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230687, 'name': ReconfigVM_Task, 'duration_secs': 0.313537} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.262074] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Reconfigured VM instance instance-00000012 to attach disk [datastore2] b23b4e67-67a6-4117-9a62-3c0dcd25e828/b23b4e67-67a6-4117-9a62-3c0dcd25e828.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.262074] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4be9a4af-00f5-4318-86a2-f0f714782ea9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.267455] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Created folder: Project (64c4d90bd33e4c82be86c8c33a1f6826) in parent group-v836999. [ 756.267455] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Creating folder: Instances. Parent ref: group-v837063. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 756.267455] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b401b0e5-934f-4fef-b9e4-9dbe1486cc49 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.277095] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 756.277095] env[70013]: value = "task-4230689" [ 756.277095] env[70013]: _type = "Task" [ 756.277095] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.287557] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Created folder: Instances in parent group-v837063. [ 756.287877] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 756.288969] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 756.291974] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3754c818-4030-422f-8e14-aee5c70423d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.313323] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230689, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.319133] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 756.319133] env[70013]: value = "task-4230691" [ 756.319133] env[70013]: _type = "Task" [ 756.319133] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.331967] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230691, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.501929] env[70013]: DEBUG nova.network.neutron [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Updated VIF entry in instance network info cache for port 47970e5c-2dab-40b7-abcc-f2bc90bcfad7. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 756.503334] env[70013]: DEBUG nova.network.neutron [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Updating instance_info_cache with network_info: [{"id": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "address": "fa:16:3e:03:9c:38", "network": {"id": "6055f3ca-7760-404c-8618-7f0e0cdc4237", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1361122985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eb984e7011845a1b0b848d8a17c6990", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47970e5c-2d", "ovs_interfaceid": "47970e5c-2dab-40b7-abcc-f2bc90bcfad7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.504555] env[70013]: DEBUG nova.network.neutron [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.561958] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Acquiring lock "566c6b47-34ce-4d53-baad-f31bf08ad3a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.562689] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Lock "566c6b47-34ce-4d53-baad-f31bf08ad3a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.689351] env[70013]: DEBUG nova.compute.manager [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Received event network-vif-plugged-5708c0e5-6147-4b18-b294-885738a29578 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 756.689351] env[70013]: DEBUG oslo_concurrency.lockutils [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] Acquiring lock "3e938bb3-54a6-49f8-809d-d5aee0349eae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.689351] env[70013]: DEBUG oslo_concurrency.lockutils [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.689726] env[70013]: DEBUG oslo_concurrency.lockutils [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.689726] env[70013]: DEBUG nova.compute.manager [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] No waiting events found dispatching network-vif-plugged-5708c0e5-6147-4b18-b294-885738a29578 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 756.689841] env[70013]: WARNING nova.compute.manager [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Received unexpected event network-vif-plugged-5708c0e5-6147-4b18-b294-885738a29578 for instance with vm_state building and task_state spawning. [ 756.690022] env[70013]: DEBUG nova.compute.manager [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Received event network-changed-5708c0e5-6147-4b18-b294-885738a29578 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 756.690182] env[70013]: DEBUG nova.compute.manager [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Refreshing instance network info cache due to event network-changed-5708c0e5-6147-4b18-b294-885738a29578. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 756.690370] env[70013]: DEBUG oslo_concurrency.lockutils [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] Acquiring lock "refresh_cache-3e938bb3-54a6-49f8-809d-d5aee0349eae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.690508] env[70013]: DEBUG oslo_concurrency.lockutils [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] Acquired lock "refresh_cache-3e938bb3-54a6-49f8-809d-d5aee0349eae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.690674] env[70013]: DEBUG nova.network.neutron [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Refreshing network info cache for port 5708c0e5-6147-4b18-b294-885738a29578 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 756.793765] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230689, 'name': Rename_Task, 'duration_secs': 0.186808} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.794083] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.794339] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-faf4abd5-abe0-424d-af2b-2217733ceefd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.803448] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 756.803448] env[70013]: value = "task-4230692" [ 756.803448] env[70013]: _type = "Task" [ 756.803448] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.814882] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.834517] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230691, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.868682] env[70013]: DEBUG nova.compute.manager [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received event network-vif-plugged-3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 756.869154] env[70013]: DEBUG oslo_concurrency.lockutils [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] Acquiring lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.869260] env[70013]: DEBUG oslo_concurrency.lockutils [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.872236] env[70013]: DEBUG oslo_concurrency.lockutils [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.872236] env[70013]: DEBUG nova.compute.manager [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] No waiting events found dispatching network-vif-plugged-3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 756.872236] env[70013]: WARNING nova.compute.manager [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received unexpected event network-vif-plugged-3ddab8d9-c5bd-4de2-994d-621e5e86cc88 for instance with vm_state building and task_state spawning. [ 756.872236] env[70013]: DEBUG nova.compute.manager [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received event network-changed-3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 756.872236] env[70013]: DEBUG nova.compute.manager [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Refreshing instance network info cache due to event network-changed-3ddab8d9-c5bd-4de2-994d-621e5e86cc88. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 756.872574] env[70013]: DEBUG oslo_concurrency.lockutils [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] Acquiring lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.007510] env[70013]: DEBUG oslo_concurrency.lockutils [req-95caf053-7251-48c0-9fbc-a3ec43fd5c5f req-33830b45-170e-4c22-9f00-7c329226083c service nova] Releasing lock "refresh_cache-b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.009579] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.010468] env[70013]: DEBUG nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Instance network_info: |[{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 757.011121] env[70013]: DEBUG oslo_concurrency.lockutils [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] Acquired lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.012229] env[70013]: DEBUG nova.network.neutron [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Refreshing network info cache for port 3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.014354] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:6e:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ddab8d9-c5bd-4de2-994d-621e5e86cc88', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.022602] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Creating folder: Project (52386d3aa84b4d3e8376dec7544fae6a). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.025009] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a71d0502-a3ea-488f-a609-15ee99fa7a3d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.038590] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Created folder: Project (52386d3aa84b4d3e8376dec7544fae6a) in parent group-v836999. [ 757.038590] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Creating folder: Instances. Parent ref: group-v837066. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.040544] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96add2fe-5670-46cc-8875-00c32bde7278 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.052653] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Created folder: Instances in parent group-v837066. [ 757.052987] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 757.053856] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 757.053856] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f76943f5-c8a9-424c-a4ec-9660fcf13d45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.082930] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.082930] env[70013]: value = "task-4230695" [ 757.082930] env[70013]: _type = "Task" [ 757.082930] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.094991] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230695, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.148516] env[70013]: DEBUG nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 757.180560] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 757.180825] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.181181] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 757.182193] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.182193] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 757.182193] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 757.182193] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 757.182193] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 757.182586] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 757.182586] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 757.182586] env[70013]: DEBUG nova.virt.hardware [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 757.183383] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ec4771-61ae-4131-a40c-18ab522fa78e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.201263] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef37f16-0f37-4197-97bd-c6b61a4f5bf1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.218864] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.227037] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Creating folder: Project (fb326ee065144430bff4061cdd821b40). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.227037] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82d252ae-2fc0-4f89-b089-57ce8fde2a3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.238691] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Created folder: Project (fb326ee065144430bff4061cdd821b40) in parent group-v836999. [ 757.238918] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Creating folder: Instances. Parent ref: group-v837069. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.239192] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85aa6785-e983-4d10-93de-7d5565a41b58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.250526] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Created folder: Instances in parent group-v837069. [ 757.251025] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 757.251890] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 757.251890] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e97b073-0e28-445c-a9fb-49145657e740 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.278123] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.278123] env[70013]: value = "task-4230698" [ 757.278123] env[70013]: _type = "Task" [ 757.278123] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.290616] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230698, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.315675] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230692, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.337023] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230691, 'name': CreateVM_Task, 'duration_secs': 0.660291} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.339259] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 757.339840] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.340247] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.340679] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 757.340900] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15fd14de-5e00-48e3-ad39-bfa77406dc78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.351716] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 757.351716] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5278dc7d-980b-36f7-48a6-4a1e03540648" [ 757.351716] env[70013]: _type = "Task" [ 757.351716] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.362433] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5278dc7d-980b-36f7-48a6-4a1e03540648, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.514271] env[70013]: DEBUG nova.network.neutron [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Updated VIF entry in instance network info cache for port 5708c0e5-6147-4b18-b294-885738a29578. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 757.514271] env[70013]: DEBUG nova.network.neutron [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Updating instance_info_cache with network_info: [{"id": "5708c0e5-6147-4b18-b294-885738a29578", "address": "fa:16:3e:5b:99:3b", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5708c0e5-61", "ovs_interfaceid": "5708c0e5-6147-4b18-b294-885738a29578", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.594225] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230695, 'name': CreateVM_Task, 'duration_secs': 0.493853} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.597863] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 757.599183] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.714292] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca06f9f-01e6-47f8-b079-f985405ec232 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.725183] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462fff8d-83ef-4e09-90a5-8266d513f21a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.763873] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea42b44-cdbf-4f4d-adfd-b94898bc8490 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.773865] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d4668d-32cf-4c13-80c0-2e4df980de10 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.779083] env[70013]: DEBUG nova.network.neutron [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updated VIF entry in instance network info cache for port 3ddab8d9-c5bd-4de2-994d-621e5e86cc88. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 757.779083] env[70013]: DEBUG nova.network.neutron [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.791683] env[70013]: DEBUG nova.compute.provider_tree [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.800644] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230698, 'name': CreateVM_Task, 'duration_secs': 0.324001} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.800644] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 757.800835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.801032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.801345] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 757.803010] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9dc4e89-fed3-4cd5-afe6-094536d5bf26 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.816065] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 757.816065] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ad8d50-bc28-794c-9045-6d4714b81d60" [ 757.816065] env[70013]: _type = "Task" [ 757.816065] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.816438] env[70013]: DEBUG oslo_vmware.api [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230692, 'name': PowerOnVM_Task, 'duration_secs': 0.636289} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.816845] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 757.817081] env[70013]: INFO nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Took 12.77 seconds to spawn the instance on the hypervisor. [ 757.817293] env[70013]: DEBUG nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 757.821597] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac22dd1e-2355-4044-a684-73772441e83f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.835201] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ad8d50-bc28-794c-9045-6d4714b81d60, 'name': SearchDatastore_Task} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.835723] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.835957] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.836209] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.837060] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.837060] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.837060] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed5c14e8-eec3-4fc2-95e1-ca307d80e5a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.849427] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.849623] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.850396] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a4629ef-8e4d-4581-9079-78b54b47c5e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.862824] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5278dc7d-980b-36f7-48a6-4a1e03540648, 'name': SearchDatastore_Task, 'duration_secs': 0.014578} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.865136] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.865136] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.865136] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.865136] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.865427] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.865472] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 757.865472] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52367988-7071-149f-0dd0-d052cd615531" [ 757.865472] env[70013]: _type = "Task" [ 757.865472] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.865651] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.866742] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 757.866742] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95e07c7a-03c2-45d8-affd-742b9e7dc86d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.868247] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a225f2d-6267-4fcf-a500-d0d4225f051c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.877309] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 757.877309] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52772038-b705-4cf2-22b0-494ae1d24692" [ 757.877309] env[70013]: _type = "Task" [ 757.877309] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.882029] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52367988-7071-149f-0dd0-d052cd615531, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.886114] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.886318] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.887579] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-265bda7a-9ccb-4f12-880e-3a0c4a03af9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.897452] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52772038-b705-4cf2-22b0-494ae1d24692, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.898516] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 757.898516] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a915aa-bb7c-5f53-be80-67792f7182bd" [ 757.898516] env[70013]: _type = "Task" [ 757.898516] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.910824] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a915aa-bb7c-5f53-be80-67792f7182bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.017389] env[70013]: DEBUG oslo_concurrency.lockutils [req-5ee45a1c-98d3-4ea8-8769-9ba148742cb5 req-c8c96120-92be-432e-b5fc-88de5a22508b service nova] Releasing lock "refresh_cache-3e938bb3-54a6-49f8-809d-d5aee0349eae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.285674] env[70013]: DEBUG oslo_concurrency.lockutils [req-680e9e89-9163-478a-b330-984a755e9892 req-32e6aa5b-261a-4df6-8e7c-0d05df5a8b5f service nova] Releasing lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.295574] env[70013]: DEBUG nova.scheduler.client.report [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 758.345670] env[70013]: INFO nova.compute.manager [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Took 35.52 seconds to build instance. [ 758.380102] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52367988-7071-149f-0dd0-d052cd615531, 'name': SearchDatastore_Task, 'duration_secs': 0.019467} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.381430] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d9ebae6-14c7-4b59-9816-5c9493ff31d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.398917] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52772038-b705-4cf2-22b0-494ae1d24692, 'name': SearchDatastore_Task, 'duration_secs': 0.025199} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.401235] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.401816] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.402523] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.402937] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 758.402937] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52da46e0-cf77-159f-56de-57ab398c2284" [ 758.402937] env[70013]: _type = "Task" [ 758.402937] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.418055] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52da46e0-cf77-159f-56de-57ab398c2284, 'name': SearchDatastore_Task, 'duration_secs': 0.011768} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.419106] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.419388] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 062eff58-95fe-4c9b-a586-7e7434c77adf/062eff58-95fe-4c9b-a586-7e7434c77adf.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 758.419680] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a915aa-bb7c-5f53-be80-67792f7182bd, 'name': SearchDatastore_Task, 'duration_secs': 0.017628} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.419881] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c80e6191-758a-468d-8b8c-6be534c68702 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.423489] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d90d329a-f9dc-476b-b78d-6e9eacef6187 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.428682] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 758.428682] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529a62d0-2f96-0148-6da7-c63bfd0e5e94" [ 758.428682] env[70013]: _type = "Task" [ 758.428682] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.433266] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 758.433266] env[70013]: value = "task-4230700" [ 758.433266] env[70013]: _type = "Task" [ 758.433266] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.440378] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529a62d0-2f96-0148-6da7-c63bfd0e5e94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.446137] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230700, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.752049] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "63c3b57c-022a-4eee-b215-6dd89142e659" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.752413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.753027] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.753027] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.753027] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.755581] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "3b29c99a-0d55-40b1-a155-199b1ced2146" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.755886] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.756081] env[70013]: DEBUG nova.compute.manager [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Going to confirm migration 1 {{(pid=70013) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 758.757991] env[70013]: INFO nova.compute.manager [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Terminating instance [ 758.805015] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.680s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.805581] env[70013]: DEBUG nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 758.808395] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.382s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.808591] env[70013]: DEBUG nova.objects.instance [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 758.849827] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6407ee1-70d5-4c39-b821-7b1e64022e29 tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.778s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.952506] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529a62d0-2f96-0148-6da7-c63bfd0e5e94, 'name': SearchDatastore_Task, 'duration_secs': 0.014315} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.953518] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230700, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.953791] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.954072] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 3e938bb3-54a6-49f8-809d-d5aee0349eae/3e938bb3-54a6-49f8-809d-d5aee0349eae.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 758.954369] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.954550] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.954769] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2b55bd0-6baa-44c0-a268-9deec134c3cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.957517] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5d36628-e306-4dcf-8f76-5c87815a6ff6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.967017] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 758.967017] env[70013]: value = "task-4230701" [ 758.967017] env[70013]: _type = "Task" [ 758.967017] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.976853] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.984668] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.984668] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9181ac24-ebe4-472f-ba3e-516c9d00a717 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.990968] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.995062] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 758.995062] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d9ab94-1b03-aa15-f91f-52e39f07ae51" [ 758.995062] env[70013]: _type = "Task" [ 758.995062] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.008278] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d9ab94-1b03-aa15-f91f-52e39f07ae51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.180407] env[70013]: DEBUG nova.compute.manager [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 759.182262] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22886ee3-ee44-4651-864a-6afe156bbc48 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.270237] env[70013]: DEBUG nova.compute.manager [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 759.274020] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.274020] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66da3bf0-fc5e-4e3e-a340-c43ba0387a2f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.288861] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 759.289603] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b13c04d-d5c5-4556-85e8-54989da83525 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.320922] env[70013]: DEBUG nova.compute.utils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 759.333920] env[70013]: DEBUG nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 759.334192] env[70013]: DEBUG nova.network.neutron [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.353744] env[70013]: DEBUG nova.compute.manager [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 759.363024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 759.363024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 759.363024] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleting the datastore file [datastore2] 63c3b57c-022a-4eee-b215-6dd89142e659 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.363366] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-052ff0d8-8911-4cf3-8605-2aafefa30dfb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.373913] env[70013]: DEBUG oslo_vmware.api [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 759.373913] env[70013]: value = "task-4230703" [ 759.373913] env[70013]: _type = "Task" [ 759.373913] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.388913] env[70013]: DEBUG oslo_vmware.api [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.452020] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230700, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.484376] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.506703] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d9ab94-1b03-aa15-f91f-52e39f07ae51, 'name': SearchDatastore_Task, 'duration_secs': 0.012714} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.507975] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d3d8d70-da16-4795-852c-7afb3d486b82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.515312] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 759.515312] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a1c84f-75f9-51ba-686f-734561364705" [ 759.515312] env[70013]: _type = "Task" [ 759.515312] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.526674] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a1c84f-75f9-51ba-686f-734561364705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.536034] env[70013]: DEBUG nova.policy [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfa9f992483942f689f616b60a27b00d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0896b76438d9456fb1b9fa2f39d0decc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 759.554244] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.555427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.556940] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.557400] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.557501] env[70013]: DEBUG nova.network.neutron [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.557780] env[70013]: DEBUG nova.objects.instance [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lazy-loading 'info_cache' on Instance uuid 3b29c99a-0d55-40b1-a155-199b1ced2146 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 759.697848] env[70013]: INFO nova.compute.manager [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] instance snapshotting [ 759.700792] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c329ad12-a100-4032-a466-3ec8db416a3a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.723776] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6163c736-9b7f-4df8-9973-91a50b60d4ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.838087] env[70013]: DEBUG nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 759.842615] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab692463-bf35-4ab8-b6cc-4c3da245b405 tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.034s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.843878] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.624s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.845124] env[70013]: DEBUG nova.objects.instance [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lazy-loading 'resources' on Instance uuid 225c02d6-43df-4260-a428-14a6b1ed32cc {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 759.893104] env[70013]: DEBUG oslo_vmware.api [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.370155} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.894748] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.894748] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.894748] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.894748] env[70013]: INFO nova.compute.manager [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Took 0.62 seconds to destroy the instance on the hypervisor. [ 759.894748] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 759.895237] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.895475] env[70013]: DEBUG nova.compute.manager [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 759.895572] env[70013]: DEBUG nova.network.neutron [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.950747] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230700, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.982092] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.030642] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a1c84f-75f9-51ba-686f-734561364705, 'name': SearchDatastore_Task, 'duration_secs': 0.021628} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.031342] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.031342] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9a0ecd6c-5853-446d-8f41-b2ee51e5259a/9a0ecd6c-5853-446d-8f41-b2ee51e5259a.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 760.031577] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0279b5fe-69b5-4657-bec5-60926dbf9ddb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.041847] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 760.041847] env[70013]: value = "task-4230705" [ 760.041847] env[70013]: _type = "Task" [ 760.041847] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.053913] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.239407] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 760.240028] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-afeea777-2a71-4817-92e7-bf1483056bad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.250252] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 760.250252] env[70013]: value = "task-4230706" [ 760.250252] env[70013]: _type = "Task" [ 760.250252] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.263435] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230706, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.446566] env[70013]: DEBUG nova.network.neutron [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Successfully created port: 3f3abd39-894f-40ce-bea2-c815ae8ed3c3 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.462333] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230700, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.539547} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.463950] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 062eff58-95fe-4c9b-a586-7e7434c77adf/062eff58-95fe-4c9b-a586-7e7434c77adf.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 760.464265] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 760.464587] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24af546e-908e-4e94-9541-db3c5cf503ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.477779] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 760.477779] env[70013]: value = "task-4230707" [ 760.477779] env[70013]: _type = "Task" [ 760.477779] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.488397] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230701, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.502685] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230707, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.553920] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230705, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.770636] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230706, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.860193] env[70013]: DEBUG nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 760.905414] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 760.905670] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.905832] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 760.906028] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.906212] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 760.906367] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 760.906578] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 760.906826] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 760.907035] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 760.907211] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 760.907388] env[70013]: DEBUG nova.virt.hardware [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 760.908671] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d72eadd-0263-4d65-9006-8653fb3e950e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.924478] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e75e5c-478d-4517-b2a2-36fc1642bc84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.973618] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8841104e-7799-4f4e-9f6e-3c604367c397 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.990221] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230701, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.52046} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.991736] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9590226d-5a84-48a6-9efe-81fba2d7240c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.995236] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 3e938bb3-54a6-49f8-809d-d5aee0349eae/3e938bb3-54a6-49f8-809d-d5aee0349eae.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 760.995605] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 760.998614] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eadd87c2-ae10-4e48-99b6-ad90d7e1b367 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.000924] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230707, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073258} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.001626] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.002760] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fda94f8-be5d-4b2c-8fe2-fb77afa3d9e9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.040677] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594c9075-91b5-4103-9eb6-eca985ee983f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.043627] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 761.043627] env[70013]: value = "task-4230708" [ 761.043627] env[70013]: _type = "Task" [ 761.043627] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.060432] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Reconfiguring VM instance instance-00000015 to attach disk [datastore2] 062eff58-95fe-4c9b-a586-7e7434c77adf/062eff58-95fe-4c9b-a586-7e7434c77adf.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.067460] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a8c368e-5f37-4fa7-8947-b63ab7b3b72f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.086948] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05fc109d-db93-437f-b396-3477ce73a0ed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.098320] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 761.098320] env[70013]: value = "task-4230709" [ 761.098320] env[70013]: _type = "Task" [ 761.098320] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.098479] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07478} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.098583] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.880223} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.099340] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.099620] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9a0ecd6c-5853-446d-8f41-b2ee51e5259a/9a0ecd6c-5853-446d-8f41-b2ee51e5259a.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 761.100169] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 761.104868] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fa95dc-f0c4-4538-8a7e-7ed3e72171ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.106478] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d7d3570-09fe-4d6f-b55d-e88b641423a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.119177] env[70013]: DEBUG nova.compute.provider_tree [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.128426] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.148063] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] 3e938bb3-54a6-49f8-809d-d5aee0349eae/3e938bb3-54a6-49f8-809d-d5aee0349eae.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.148524] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 761.148524] env[70013]: value = "task-4230710" [ 761.148524] env[70013]: _type = "Task" [ 761.148524] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.149464] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd3856d9-a3a4-4ff0-990f-3bc0ab00a280 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.168719] env[70013]: DEBUG nova.network.neutron [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.188093] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 761.188093] env[70013]: value = "task-4230711" [ 761.188093] env[70013]: _type = "Task" [ 761.188093] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.190023] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230710, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.206222] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230711, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.263244] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230706, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.503994] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "77c12460-9cfa-41c9-a210-238a470d9ccd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 761.504221] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.517718] env[70013]: DEBUG nova.network.neutron [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance_info_cache with network_info: [{"id": "b402fce4-7392-4139-8de5-7620820e0814", "address": "fa:16:3e:44:ef:b6", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb402fce4-73", "ovs_interfaceid": "b402fce4-7392-4139-8de5-7620820e0814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.617452] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230709, 'name': ReconfigVM_Task, 'duration_secs': 0.33792} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.617891] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Reconfigured VM instance instance-00000015 to attach disk [datastore2] 062eff58-95fe-4c9b-a586-7e7434c77adf/062eff58-95fe-4c9b-a586-7e7434c77adf.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.618671] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fa5a1a4-2c83-48e6-8794-559b30762110 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.623406] env[70013]: DEBUG nova.scheduler.client.report [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.630534] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 761.630534] env[70013]: value = "task-4230712" [ 761.630534] env[70013]: _type = "Task" [ 761.630534] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.646961] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230712, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.683502] env[70013]: INFO nova.compute.manager [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Took 1.79 seconds to deallocate network for instance. [ 761.683924] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230710, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102138} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.685875] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.691143] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a6195d-6d15-4f59-82c2-706770578130 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.725938] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] 9a0ecd6c-5853-446d-8f41-b2ee51e5259a/9a0ecd6c-5853-446d-8f41-b2ee51e5259a.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.733045] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f467fa1b-65b7-4410-b3ef-ccbaf86b93f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.751808] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230711, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.759544] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 761.759544] env[70013]: value = "task-4230713" [ 761.759544] env[70013]: _type = "Task" [ 761.759544] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.765221] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230706, 'name': CreateSnapshot_Task, 'duration_secs': 1.242343} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.767680] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 761.768960] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0275e21-01d1-48b3-bba6-cc092e41bb29 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.778693] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230713, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.001462] env[70013]: DEBUG nova.compute.manager [req-7cf2d5d1-9f84-420f-b619-29a164508746 req-52857a76-a328-4098-adc4-709273d0a902 service nova] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Received event network-vif-deleted-a4ff3dbe-e842-4e4a-939e-e3f429ebad04 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 762.021233] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "refresh_cache-3b29c99a-0d55-40b1-a155-199b1ced2146" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.021502] env[70013]: DEBUG nova.objects.instance [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lazy-loading 'migration_context' on Instance uuid 3b29c99a-0d55-40b1-a155-199b1ced2146 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.129042] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.285s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.132461] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.494s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.134850] env[70013]: INFO nova.compute.claims [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.155249] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230712, 'name': Rename_Task, 'duration_secs': 0.161388} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.155616] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.155882] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd7c1831-e80a-4a5b-a076-d910b2b4a060 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.161358] env[70013]: INFO nova.scheduler.client.report [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Deleted allocations for instance 225c02d6-43df-4260-a428-14a6b1ed32cc [ 762.168344] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 762.168344] env[70013]: value = "task-4230714" [ 762.168344] env[70013]: _type = "Task" [ 762.168344] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.180595] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.199224] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.213837] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230711, 'name': ReconfigVM_Task, 'duration_secs': 0.542397} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.213837] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Reconfigured VM instance instance-00000013 to attach disk [datastore1] 3e938bb3-54a6-49f8-809d-d5aee0349eae/3e938bb3-54a6-49f8-809d-d5aee0349eae.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 762.213837] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37b806c8-a87c-4df0-9a38-9551830f5fde {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.223090] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 762.223090] env[70013]: value = "task-4230715" [ 762.223090] env[70013]: _type = "Task" [ 762.223090] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.235848] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230715, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.274445] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230713, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.292762] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 762.293120] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f2d58111-7d70-4f7f-bf47-6adf6a0fa845 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.303436] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 762.303436] env[70013]: value = "task-4230716" [ 762.303436] env[70013]: _type = "Task" [ 762.303436] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.316835] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230716, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.526454] env[70013]: DEBUG nova.objects.base [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Object Instance<3b29c99a-0d55-40b1-a155-199b1ced2146> lazy-loaded attributes: info_cache,migration_context {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 762.527494] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd546d4d-163f-4f66-a760-7ff42eecb857 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.552440] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-359d4934-87d9-4182-9e39-1bf8db783ee0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.560088] env[70013]: DEBUG oslo_vmware.api [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 762.560088] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52120e41-6890-8d44-7e8e-1c24b223f9ef" [ 762.560088] env[70013]: _type = "Task" [ 762.560088] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.572647] env[70013]: DEBUG oslo_vmware.api [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52120e41-6890-8d44-7e8e-1c24b223f9ef, 'name': SearchDatastore_Task, 'duration_secs': 0.00931} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.572797] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.675138] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d34ac467-7773-4252-abd2-a2018dc26168 tempest-AttachInterfacesV270Test-841799969 tempest-AttachInterfacesV270Test-841799969-project-member] Lock "225c02d6-43df-4260-a428-14a6b1ed32cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.277s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.685935] env[70013]: DEBUG oslo_vmware.api [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230714, 'name': PowerOnVM_Task, 'duration_secs': 0.518532} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.686680] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 762.686912] env[70013]: INFO nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Took 5.54 seconds to spawn the instance on the hypervisor. [ 762.687132] env[70013]: DEBUG nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 762.687940] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c1e343-31f4-41d8-9017-d0b1e5988ad2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.736387] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230715, 'name': Rename_Task, 'duration_secs': 0.275635} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.738323] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.738422] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-541be51e-416c-4bbc-8888-2b13b8e85102 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.749536] env[70013]: DEBUG nova.network.neutron [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Successfully updated port: 3f3abd39-894f-40ce-bea2-c815ae8ed3c3 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 762.754915] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 762.754915] env[70013]: value = "task-4230717" [ 762.754915] env[70013]: _type = "Task" [ 762.754915] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.767167] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230717, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.779340] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230713, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.815304] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230716, 'name': CloneVM_Task} progress is 93%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.946057] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.946333] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.214366] env[70013]: INFO nova.compute.manager [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Took 32.81 seconds to build instance. [ 763.256682] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-09544922-6dc4-48b5-8cfd-e91e7f74c13f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.256910] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-09544922-6dc4-48b5-8cfd-e91e7f74c13f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.257119] env[70013]: DEBUG nova.network.neutron [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.287619] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230713, 'name': ReconfigVM_Task, 'duration_secs': 1.430719} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.292607] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Reconfigured VM instance instance-00000014 to attach disk [datastore1] 9a0ecd6c-5853-446d-8f41-b2ee51e5259a/9a0ecd6c-5853-446d-8f41-b2ee51e5259a.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.294414] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230717, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.299889] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17fe0671-40fb-4c9b-b285-67faa5cbe36c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.308682] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 763.308682] env[70013]: value = "task-4230718" [ 763.308682] env[70013]: _type = "Task" [ 763.308682] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.329485] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230716, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.329911] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230718, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.423738] env[70013]: DEBUG nova.compute.manager [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Received event network-changed-7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 763.424314] env[70013]: DEBUG nova.compute.manager [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Refreshing instance network info cache due to event network-changed-7634504a-a16c-44ea-9a82-e8f88bcc65ea. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 763.424498] env[70013]: DEBUG oslo_concurrency.lockutils [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] Acquiring lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.424739] env[70013]: DEBUG oslo_concurrency.lockutils [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] Acquired lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.424877] env[70013]: DEBUG nova.network.neutron [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Refreshing network info cache for port 7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 763.637873] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc52d8f9-a699-4f69-96ac-1ac951f8b31c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.647729] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a9abca-73a2-4d62-8eee-c34ed67546e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.687308] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5909854d-523d-4a64-b463-83d58c5309da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.696481] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc91b95-a665-42cb-9f2c-7dd44615b071 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.712756] env[70013]: DEBUG nova.compute.provider_tree [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.722012] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33e9edcf-a190-4b0f-99aa-27c20e5d9cd5 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "062eff58-95fe-4c9b-a586-7e7434c77adf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.191s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.775029] env[70013]: DEBUG oslo_vmware.api [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230717, 'name': PowerOnVM_Task, 'duration_secs': 0.77244} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.778306] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.778306] env[70013]: INFO nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Took 14.29 seconds to spawn the instance on the hypervisor. [ 763.778306] env[70013]: DEBUG nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 763.778306] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a24787-df42-49db-87aa-094bffb497ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.814177] env[70013]: DEBUG nova.network.neutron [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.827692] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230716, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.831279] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230718, 'name': Rename_Task, 'duration_secs': 0.229052} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.831701] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 763.831946] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-971f28f0-086e-4805-bf6d-4f447c23526e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.841741] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 763.841741] env[70013]: value = "task-4230719" [ 763.841741] env[70013]: _type = "Task" [ 763.841741] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.852816] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.979034] env[70013]: DEBUG nova.network.neutron [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Updating instance_info_cache with network_info: [{"id": "3f3abd39-894f-40ce-bea2-c815ae8ed3c3", "address": "fa:16:3e:bf:d0:c8", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f3abd39-89", "ovs_interfaceid": "3f3abd39-894f-40ce-bea2-c815ae8ed3c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.219570] env[70013]: DEBUG nova.scheduler.client.report [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.229118] env[70013]: DEBUG nova.compute.manager [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 764.304469] env[70013]: INFO nova.compute.manager [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Took 36.54 seconds to build instance. [ 764.322739] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230716, 'name': CloneVM_Task, 'duration_secs': 2.002906} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.325741] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Created linked-clone VM from snapshot [ 764.327127] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb18c0cd-30b6-417f-8931-2efe4c9b3c46 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.338830] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Uploading image 00d5405f-9c08-411a-9d32-f09745461ee9 {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 764.354040] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230719, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.379062] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 764.379062] env[70013]: value = "vm-837073" [ 764.379062] env[70013]: _type = "VirtualMachine" [ 764.379062] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 764.379330] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fce42d34-2a05-4565-880e-af2c5bdb4438 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.389304] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lease: (returnval){ [ 764.389304] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522d0faf-82a9-ceaf-7e56-a46efa9be089" [ 764.389304] env[70013]: _type = "HttpNfcLease" [ 764.389304] env[70013]: } obtained for exporting VM: (result){ [ 764.389304] env[70013]: value = "vm-837073" [ 764.389304] env[70013]: _type = "VirtualMachine" [ 764.389304] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 764.389867] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the lease: (returnval){ [ 764.389867] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522d0faf-82a9-ceaf-7e56-a46efa9be089" [ 764.389867] env[70013]: _type = "HttpNfcLease" [ 764.389867] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 764.397417] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 764.397417] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522d0faf-82a9-ceaf-7e56-a46efa9be089" [ 764.397417] env[70013]: _type = "HttpNfcLease" [ 764.397417] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 764.435480] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquiring lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 764.435738] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.482045] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-09544922-6dc4-48b5-8cfd-e91e7f74c13f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.483011] env[70013]: DEBUG nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Instance network_info: |[{"id": "3f3abd39-894f-40ce-bea2-c815ae8ed3c3", "address": "fa:16:3e:bf:d0:c8", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f3abd39-89", "ovs_interfaceid": "3f3abd39-894f-40ce-bea2-c815ae8ed3c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 764.483178] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:d0:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac7039c0-3374-4c08-87fc-af2449b48b02', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f3abd39-894f-40ce-bea2-c815ae8ed3c3', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 764.491807] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating folder: Project (0896b76438d9456fb1b9fa2f39d0decc). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 764.492149] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a0f6179-3b97-4a1e-ab5c-04fa53036d57 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.504478] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created folder: Project (0896b76438d9456fb1b9fa2f39d0decc) in parent group-v836999. [ 764.504566] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating folder: Instances. Parent ref: group-v837074. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 764.504768] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a71c0339-7dc2-41bd-a79b-091d7e461c20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.518135] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created folder: Instances in parent group-v837074. [ 764.518341] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 764.518561] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 764.518723] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5dc25c3f-a101-4306-8c74-9562194e4e5c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.539784] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 764.539784] env[70013]: value = "task-4230723" [ 764.539784] env[70013]: _type = "Task" [ 764.539784] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.550141] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230723, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.734518] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.735139] env[70013]: DEBUG nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 764.742521] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.410s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.742850] env[70013]: DEBUG nova.objects.instance [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lazy-loading 'resources' on Instance uuid 64308d7c-b63b-4fa6-b235-2f0586a888cf {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 764.770121] env[70013]: DEBUG nova.network.neutron [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Updated VIF entry in instance network info cache for port 7634504a-a16c-44ea-9a82-e8f88bcc65ea. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 764.771828] env[70013]: DEBUG nova.network.neutron [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Updating instance_info_cache with network_info: [{"id": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "address": "fa:16:3e:79:3b:a7", "network": {"id": "dabcb531-f742-4f77-a8b8-97269090c723", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1602634587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c266552bee9a48dfa99379521fb8efdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7634504a-a1", "ovs_interfaceid": "7634504a-a16c-44ea-9a82-e8f88bcc65ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.774956] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 764.807284] env[70013]: DEBUG oslo_concurrency.lockutils [None req-292c5b2d-d377-4f51-8fc3-0e9a49b629e5 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.050s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.855098] env[70013]: DEBUG oslo_vmware.api [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230719, 'name': PowerOnVM_Task, 'duration_secs': 0.518976} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.858329] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 764.858329] env[70013]: INFO nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Took 12.65 seconds to spawn the instance on the hypervisor. [ 764.858329] env[70013]: DEBUG nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 764.858579] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8ff210-a532-4baa-9023-61d534bea32f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.901770] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 764.901770] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522d0faf-82a9-ceaf-7e56-a46efa9be089" [ 764.901770] env[70013]: _type = "HttpNfcLease" [ 764.901770] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 764.902107] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 764.902107] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522d0faf-82a9-ceaf-7e56-a46efa9be089" [ 764.902107] env[70013]: _type = "HttpNfcLease" [ 764.902107] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 764.902861] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025f8ead-4602-463a-b7cf-07609972c539 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.912472] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279d883-6aae-cd9e-4730-12790f47f3f6/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 764.912786] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279d883-6aae-cd9e-4730-12790f47f3f6/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 765.051398] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230723, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.065906] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a45053eb-cbe1-4e54-806c-e88517645e43 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.243180] env[70013]: DEBUG nova.compute.utils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 765.245374] env[70013]: DEBUG nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 765.245374] env[70013]: DEBUG nova.network.neutron [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.274791] env[70013]: DEBUG oslo_concurrency.lockutils [req-d1ebaab6-23e5-458b-a4c2-2b354ba8f92a req-8830dd0e-ec36-4c7a-90e4-2e799a7daa1c service nova] Releasing lock "refresh_cache-b23b4e67-67a6-4117-9a62-3c0dcd25e828" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.314495] env[70013]: DEBUG nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 765.366338] env[70013]: DEBUG nova.policy [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e2f2a7779c54a1e863e94096aec6283', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f0e36e7872c4024b587dc70b9a15ea0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 765.382639] env[70013]: INFO nova.compute.manager [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Took 36.99 seconds to build instance. [ 765.449234] env[70013]: DEBUG nova.compute.manager [None req-e76f4210-f871-435a-9d58-7ba0cee14160 tempest-ServerDiagnosticsV248Test-1970149035 tempest-ServerDiagnosticsV248Test-1970149035-project-admin] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 765.453191] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4dafe44-6f48-48fa-9612-71127107050b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.476190] env[70013]: INFO nova.compute.manager [None req-e76f4210-f871-435a-9d58-7ba0cee14160 tempest-ServerDiagnosticsV248Test-1970149035 tempest-ServerDiagnosticsV248Test-1970149035-project-admin] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Retrieving diagnostics [ 765.476190] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0a3c9d-2081-47dd-88b2-2aa019f76f6c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.559670] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230723, 'name': CreateVM_Task, 'duration_secs': 0.732918} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.561037] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 765.561340] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.564501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.564501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 765.564501] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f04edc9-7ce5-41a0-b417-c4594f364b5d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.571819] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 765.571819] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527995a0-e572-6cc6-8cac-9c51862d2dde" [ 765.571819] env[70013]: _type = "Task" [ 765.571819] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.602899] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527995a0-e572-6cc6-8cac-9c51862d2dde, 'name': SearchDatastore_Task, 'duration_secs': 0.015033} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.607654] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.608056] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 765.608421] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.608666] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.608942] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 765.610760] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8430324-0905-4523-bf37-5109bc30a607 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.622835] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 765.624452] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 765.625263] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d89f7d16-b658-48dd-99de-3cfb0ffaecab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.639912] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 765.639912] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c5730e-002e-e8c6-e141-4378fafdb8d3" [ 765.639912] env[70013]: _type = "Task" [ 765.639912] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.653549] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c5730e-002e-e8c6-e141-4378fafdb8d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.751774] env[70013]: DEBUG nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 765.849036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.887946] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eccf7839-f656-456f-a344-028e178215be tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.509s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.921652] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74d998b-92ff-4405-894b-f94cc4f84a6f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.933227] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b97501d-1750-44ec-9cd5-fa16a55d26a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.983873] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a82a7bb-382d-4539-8c4b-c691263f61dc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.993566] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e733965e-91ae-4d76-9bd3-f28d7c0c0d7f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.011228] env[70013]: DEBUG nova.compute.provider_tree [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.056319] env[70013]: DEBUG nova.compute.manager [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Received event network-vif-plugged-3f3abd39-894f-40ce-bea2-c815ae8ed3c3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 766.056431] env[70013]: DEBUG oslo_concurrency.lockutils [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] Acquiring lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.056616] env[70013]: DEBUG oslo_concurrency.lockutils [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.057031] env[70013]: DEBUG oslo_concurrency.lockutils [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.058427] env[70013]: DEBUG nova.compute.manager [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] No waiting events found dispatching network-vif-plugged-3f3abd39-894f-40ce-bea2-c815ae8ed3c3 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 766.058427] env[70013]: WARNING nova.compute.manager [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Received unexpected event network-vif-plugged-3f3abd39-894f-40ce-bea2-c815ae8ed3c3 for instance with vm_state building and task_state spawning. [ 766.058427] env[70013]: DEBUG nova.compute.manager [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Received event network-changed-3f3abd39-894f-40ce-bea2-c815ae8ed3c3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 766.058427] env[70013]: DEBUG nova.compute.manager [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Refreshing instance network info cache due to event network-changed-3f3abd39-894f-40ce-bea2-c815ae8ed3c3. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 766.058427] env[70013]: DEBUG oslo_concurrency.lockutils [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] Acquiring lock "refresh_cache-09544922-6dc4-48b5-8cfd-e91e7f74c13f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.058758] env[70013]: DEBUG oslo_concurrency.lockutils [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] Acquired lock "refresh_cache-09544922-6dc4-48b5-8cfd-e91e7f74c13f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.058758] env[70013]: DEBUG nova.network.neutron [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Refreshing network info cache for port 3f3abd39-894f-40ce-bea2-c815ae8ed3c3 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 766.155750] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c5730e-002e-e8c6-e141-4378fafdb8d3, 'name': SearchDatastore_Task, 'duration_secs': 0.022512} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.155750] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3003cc8-4551-44cf-827f-3db5e7b98770 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.163291] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 766.163291] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5235c938-cba4-c54d-a270-dee504c2f6fc" [ 766.163291] env[70013]: _type = "Task" [ 766.163291] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.173894] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5235c938-cba4-c54d-a270-dee504c2f6fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.395088] env[70013]: DEBUG nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 766.517068] env[70013]: DEBUG nova.scheduler.client.report [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 766.537058] env[70013]: DEBUG nova.network.neutron [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Successfully created port: aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.679892] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5235c938-cba4-c54d-a270-dee504c2f6fc, 'name': SearchDatastore_Task, 'duration_secs': 0.0153} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.680278] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.680861] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 09544922-6dc4-48b5-8cfd-e91e7f74c13f/09544922-6dc4-48b5-8cfd-e91e7f74c13f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 766.681228] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e470168c-4f4a-411c-ba48-40d4c58baf36 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.691088] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 766.691088] env[70013]: value = "task-4230724" [ 766.691088] env[70013]: _type = "Task" [ 766.691088] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.708684] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230724, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.768936] env[70013]: DEBUG nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 766.811362] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 766.811646] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.811832] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.812241] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.812459] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.812625] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 766.812879] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 766.813117] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 766.813323] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 766.813597] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 766.813778] env[70013]: DEBUG nova.virt.hardware [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 766.817126] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcef8924-c8fb-47ab-b676-fcf779437573 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.828722] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468dd028-072a-435c-8f4b-1d7d6ff6e5cf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.928928] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.973477] env[70013]: DEBUG nova.network.neutron [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Updated VIF entry in instance network info cache for port 3f3abd39-894f-40ce-bea2-c815ae8ed3c3. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 766.973477] env[70013]: DEBUG nova.network.neutron [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Updating instance_info_cache with network_info: [{"id": "3f3abd39-894f-40ce-bea2-c815ae8ed3c3", "address": "fa:16:3e:bf:d0:c8", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f3abd39-89", "ovs_interfaceid": "3f3abd39-894f-40ce-bea2-c815ae8ed3c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.028235] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.285s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.029573] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.229s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.029890] env[70013]: DEBUG nova.objects.instance [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lazy-loading 'resources' on Instance uuid 98a0199a-4209-4faf-adf5-7ae33b099d20 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.060103] env[70013]: INFO nova.scheduler.client.report [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Deleted allocations for instance 64308d7c-b63b-4fa6-b235-2f0586a888cf [ 767.206437] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230724, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.476777] env[70013]: DEBUG oslo_concurrency.lockutils [req-1008c468-337b-41b1-9f97-f15b7322677f req-0f460083-ca0a-4f39-9c09-20126675a78b service nova] Releasing lock "refresh_cache-09544922-6dc4-48b5-8cfd-e91e7f74c13f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.568687] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8ef75a04-d9b1-44b3-b050-96324b08d6fd tempest-ServerDiagnosticsNegativeTest-1600725784 tempest-ServerDiagnosticsNegativeTest-1600725784-project-member] Lock "64308d7c-b63b-4fa6-b235-2f0586a888cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.340s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.710298] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230724, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730094} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.710298] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 09544922-6dc4-48b5-8cfd-e91e7f74c13f/09544922-6dc4-48b5-8cfd-e91e7f74c13f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 767.710298] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 767.713177] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04ee86e5-9c56-4dbe-835f-59ef9ae97c88 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.723927] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 767.723927] env[70013]: value = "task-4230725" [ 767.723927] env[70013]: _type = "Task" [ 767.723927] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.737727] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230725, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.032719] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17ef856-bac0-455e-9665-cb6b6cdfb46d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.041688] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134ed6bb-44c3-44ed-8204-f6c7bd3d01f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.079502] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f56677b-1565-445c-b277-cce2d1d60892 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.089310] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b3d832-3312-4155-9823-bbcf777c718c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.105349] env[70013]: DEBUG nova.compute.provider_tree [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.235017] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230725, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076812} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.235427] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.236260] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03be47bb-83db-4a46-818f-7aeb60e81481 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.259567] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 09544922-6dc4-48b5-8cfd-e91e7f74c13f/09544922-6dc4-48b5-8cfd-e91e7f74c13f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.260324] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7f8317f-b9c1-44ec-ad35-758a5ded4ef4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.282599] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 768.282599] env[70013]: value = "task-4230726" [ 768.282599] env[70013]: _type = "Task" [ 768.282599] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.291657] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230726, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.549173] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "e39ec7f2-76d7-4794-9812-e4eb23a405e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.549576] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "e39ec7f2-76d7-4794-9812-e4eb23a405e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.610261] env[70013]: DEBUG nova.scheduler.client.report [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 768.697635] env[70013]: DEBUG nova.network.neutron [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Successfully updated port: aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 768.801276] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230726, 'name': ReconfigVM_Task, 'duration_secs': 0.50665} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.801276] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 09544922-6dc4-48b5-8cfd-e91e7f74c13f/09544922-6dc4-48b5-8cfd-e91e7f74c13f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 768.801276] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e561c789-1b8f-47b6-911c-ab31a17eaf65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.810287] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 768.810287] env[70013]: value = "task-4230727" [ 768.810287] env[70013]: _type = "Task" [ 768.810287] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.821532] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230727, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.974776] env[70013]: DEBUG nova.compute.manager [req-38a23196-b15a-411e-8cf3-d0f6b95ad2de req-4a9cf738-ec83-4d19-9e60-0048349fbdbc service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Received event network-vif-plugged-aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 768.974776] env[70013]: DEBUG oslo_concurrency.lockutils [req-38a23196-b15a-411e-8cf3-d0f6b95ad2de req-4a9cf738-ec83-4d19-9e60-0048349fbdbc service nova] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.974776] env[70013]: DEBUG oslo_concurrency.lockutils [req-38a23196-b15a-411e-8cf3-d0f6b95ad2de req-4a9cf738-ec83-4d19-9e60-0048349fbdbc service nova] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.976121] env[70013]: DEBUG oslo_concurrency.lockutils [req-38a23196-b15a-411e-8cf3-d0f6b95ad2de req-4a9cf738-ec83-4d19-9e60-0048349fbdbc service nova] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.976556] env[70013]: DEBUG nova.compute.manager [req-38a23196-b15a-411e-8cf3-d0f6b95ad2de req-4a9cf738-ec83-4d19-9e60-0048349fbdbc service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] No waiting events found dispatching network-vif-plugged-aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 768.977061] env[70013]: WARNING nova.compute.manager [req-38a23196-b15a-411e-8cf3-d0f6b95ad2de req-4a9cf738-ec83-4d19-9e60-0048349fbdbc service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Received unexpected event network-vif-plugged-aab77ffc-40a3-47ea-abe4-1abb8298eba0 for instance with vm_state building and task_state spawning. [ 769.100537] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.100537] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.116672] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.087s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.120051] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.643s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.124063] env[70013]: DEBUG nova.objects.instance [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lazy-loading 'resources' on Instance uuid cf8be667-da84-466b-9b7b-3bdfda6a8193 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 769.152067] env[70013]: INFO nova.scheduler.client.report [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Deleted allocations for instance 98a0199a-4209-4faf-adf5-7ae33b099d20 [ 769.202989] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.203527] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.203527] env[70013]: DEBUG nova.network.neutron [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.321618] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230727, 'name': Rename_Task, 'duration_secs': 0.199203} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.321930] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 769.322206] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a72ae916-d71b-4147-955f-fe40578bfee8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.330130] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 769.330130] env[70013]: value = "task-4230728" [ 769.330130] env[70013]: _type = "Task" [ 769.330130] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.340024] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.602536] env[70013]: DEBUG nova.compute.utils [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 769.668413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-532259ff-9e6f-40ee-8d43-6f26b062cb2b tempest-ServersAdmin275Test-577712933 tempest-ServersAdmin275Test-577712933-project-member] Lock "98a0199a-4209-4faf-adf5-7ae33b099d20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.423s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.723531] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "f1871673-e764-460b-adce-1742e7c105c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.723730] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "f1871673-e764-460b-adce-1742e7c105c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.799964] env[70013]: DEBUG nova.network.neutron [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.850966] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230728, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.111819] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.200669] env[70013]: DEBUG nova.network.neutron [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating instance_info_cache with network_info: [{"id": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "address": "fa:16:3e:5d:0d:2d", "network": {"id": "408751e1-e468-4c61-b5c6-06c709856e70", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-745309513-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f0e36e7872c4024b587dc70b9a15ea0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab77ffc-40", "ovs_interfaceid": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.282314] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286e31ec-ea0e-434e-a920-239e559bbd10 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.291247] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec3b372-295f-4a8e-b589-f1bad3b02e92 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.329406] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218a8b22-e720-44f0-a9a3-f13b5e7b707c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.346703] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce39ac8-ad39-41a7-9eb9-cd71fc4fa597 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.352393] env[70013]: DEBUG oslo_vmware.api [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230728, 'name': PowerOnVM_Task, 'duration_secs': 0.814757} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.352674] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 770.352879] env[70013]: INFO nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Took 9.49 seconds to spawn the instance on the hypervisor. [ 770.353073] env[70013]: DEBUG nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 770.354235] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0c4624-0987-4a79-aff8-a633b880a9f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.367866] env[70013]: DEBUG nova.compute.provider_tree [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.379377] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Acquiring lock "0b42f938-50aa-4576-9231-d8f3461ee94e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.380668] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Lock "0b42f938-50aa-4576-9231-d8f3461ee94e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.531606] env[70013]: DEBUG nova.compute.manager [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received event network-changed-3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 770.531606] env[70013]: DEBUG nova.compute.manager [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Refreshing instance network info cache due to event network-changed-3ddab8d9-c5bd-4de2-994d-621e5e86cc88. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 770.531606] env[70013]: DEBUG oslo_concurrency.lockutils [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] Acquiring lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.531606] env[70013]: DEBUG oslo_concurrency.lockutils [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] Acquired lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.531606] env[70013]: DEBUG nova.network.neutron [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Refreshing network info cache for port 3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 770.703941] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Releasing lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.704549] env[70013]: DEBUG nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance network_info: |[{"id": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "address": "fa:16:3e:5d:0d:2d", "network": {"id": "408751e1-e468-4c61-b5c6-06c709856e70", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-745309513-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f0e36e7872c4024b587dc70b9a15ea0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab77ffc-40", "ovs_interfaceid": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 770.705560] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:0d:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1430a695-49fb-4905-bc38-db9b869a1a9d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aab77ffc-40a3-47ea-abe4-1abb8298eba0', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.715450] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Creating folder: Project (5f0e36e7872c4024b587dc70b9a15ea0). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 770.716935] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2b00f85-0af5-456f-b6b7-4b189a960033 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.733092] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Created folder: Project (5f0e36e7872c4024b587dc70b9a15ea0) in parent group-v836999. [ 770.733330] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Creating folder: Instances. Parent ref: group-v837077. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 770.733809] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed1067cc-f7c4-4d86-ad40-bad8c4b20a27 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.746738] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Created folder: Instances in parent group-v837077. [ 770.746857] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 770.747299] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 770.747757] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b41502a6-42ef-49a5-abe0-02352757b2a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.769584] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 770.769584] env[70013]: value = "task-4230731" [ 770.769584] env[70013]: _type = "Task" [ 770.769584] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.780264] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230731, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.877235] env[70013]: DEBUG nova.scheduler.client.report [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 770.895469] env[70013]: INFO nova.compute.manager [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Took 37.22 seconds to build instance. [ 771.026266] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "352790fe-057f-45ee-aff0-549b456fd181" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.026762] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "352790fe-057f-45ee-aff0-549b456fd181" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.202838] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.203649] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.203922] env[70013]: INFO nova.compute.manager [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Attaching volume d827861d-a75d-44ae-8fdc-a94870d66776 to /dev/sdb [ 771.266018] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7bd751-37a3-4c8d-a52c-59ff07eead26 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.286577] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360f3dc8-1ddd-4b78-ba1c-f542776b6f5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.292244] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230731, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.312390] env[70013]: DEBUG nova.virt.block_device [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updating existing volume attachment record: ebd3443a-a430-4ac9-8616-b0ade72526b5 {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 771.383232] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.263s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.386459] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.473s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.388627] env[70013]: INFO nova.compute.claims [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.400213] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6fc4a87b-b8b3-4d59-a73a-0734feca339d tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.735s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.449997] env[70013]: INFO nova.scheduler.client.report [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Deleted allocations for instance cf8be667-da84-466b-9b7b-3bdfda6a8193 [ 772.374923] env[70013]: DEBUG nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 772.380760] env[70013]: DEBUG nova.compute.manager [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Received event network-changed-aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 772.380918] env[70013]: DEBUG nova.compute.manager [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Refreshing instance network info cache due to event network-changed-aab77ffc-40a3-47ea-abe4-1abb8298eba0. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 772.382776] env[70013]: DEBUG oslo_concurrency.lockutils [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] Acquiring lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.382776] env[70013]: DEBUG oslo_concurrency.lockutils [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] Acquired lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.382776] env[70013]: DEBUG nova.network.neutron [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Refreshing network info cache for port aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.383892] env[70013]: DEBUG oslo_concurrency.lockutils [None req-98274ab9-1121-4428-96ef-746064ce811d tempest-ServerShowV254Test-127190373 tempest-ServerShowV254Test-127190373-project-member] Lock "cf8be667-da84-466b-9b7b-3bdfda6a8193" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.366s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.397271] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230731, 'name': CreateVM_Task, 'duration_secs': 0.582482} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.397418] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 772.398243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.399265] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.399265] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 772.399894] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cbd6cd3-8018-4f78-82ae-05a92cd8d97f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.408738] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 772.408738] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523a42d6-90d0-2605-3b2a-300045e01348" [ 772.408738] env[70013]: _type = "Task" [ 772.408738] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.420955] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523a42d6-90d0-2605-3b2a-300045e01348, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.638713] env[70013]: DEBUG nova.network.neutron [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updated VIF entry in instance network info cache for port 3ddab8d9-c5bd-4de2-994d-621e5e86cc88. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 772.639874] env[70013]: DEBUG nova.network.neutron [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.915570] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.924042] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523a42d6-90d0-2605-3b2a-300045e01348, 'name': SearchDatastore_Task, 'duration_secs': 0.013297} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.924406] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.924776] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.924914] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.925118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.925722] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.925722] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-efe641b2-2977-4d68-ae7c-c7b67dfecd81 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.937707] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.939269] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.942161] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22e10c00-9402-42b2-9bc3-7493a4bcf320 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.954433] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 772.954433] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e70cc6-184e-efe7-c71b-143ac258f8d6" [ 772.954433] env[70013]: _type = "Task" [ 772.954433] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.965717] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e70cc6-184e-efe7-c71b-143ac258f8d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.128280] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e668e455-0474-4b05-80d7-881ae19e3a8c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "d3d5b3ff-9ac4-4c6b-b9d9-daa306275c5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.128280] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e668e455-0474-4b05-80d7-881ae19e3a8c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "d3d5b3ff-9ac4-4c6b-b9d9-daa306275c5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.148394] env[70013]: DEBUG oslo_concurrency.lockutils [req-15a486e4-984e-4bde-bf25-411f6ca3f0ec req-5d2b97db-420b-451d-8c34-ded187f9f416 service nova] Releasing lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.273503] env[70013]: DEBUG nova.network.neutron [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updated VIF entry in instance network info cache for port aab77ffc-40a3-47ea-abe4-1abb8298eba0. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 773.274074] env[70013]: DEBUG nova.network.neutron [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating instance_info_cache with network_info: [{"id": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "address": "fa:16:3e:5d:0d:2d", "network": {"id": "408751e1-e468-4c61-b5c6-06c709856e70", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-745309513-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f0e36e7872c4024b587dc70b9a15ea0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab77ffc-40", "ovs_interfaceid": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.471640] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e70cc6-184e-efe7-c71b-143ac258f8d6, 'name': SearchDatastore_Task, 'duration_secs': 0.016258} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.472478] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc88c67c-18e5-42c5-afb3-8866260e147a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.480587] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 773.480587] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c1760-daa7-1bca-06ae-d2c9fe37171f" [ 773.480587] env[70013]: _type = "Task" [ 773.480587] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.497061] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c1760-daa7-1bca-06ae-d2c9fe37171f, 'name': SearchDatastore_Task, 'duration_secs': 0.013978} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.497416] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.497697] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2c0e4326-c33a-42bb-b793-a100157b1c03/2c0e4326-c33a-42bb-b793-a100157b1c03.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.497988] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7173d390-1c07-41c6-aeb7-5b74c6e80ca5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.506768] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 773.506768] env[70013]: value = "task-4230735" [ 773.506768] env[70013]: _type = "Task" [ 773.506768] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.522513] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.547417] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60915016-8c92-4651-bfde-c4f99665a6f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.556669] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14afe286-6c02-4ea1-a3a5-153f781e9738 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.590503] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911a2bd6-f5c9-448e-b116-2c2ae9c1a565 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.599771] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1872b8b1-ae70-4ded-a8d6-5869ca87a9f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.616102] env[70013]: DEBUG nova.compute.provider_tree [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.762456] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "02ef7875-7ec9-4409-aaa5-71ed669f3780" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.763036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.782251] env[70013]: DEBUG oslo_concurrency.lockutils [req-e407f185-eeee-47a5-91a5-ebc2efff36ba req-9b5bfac1-93d8-4309-8bf9-105dc2f0a7b0 service nova] Releasing lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.799860] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "b1d844cc-dcdc-47e2-85c3-484d0862be60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.800230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.836398] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "c0e830f3-e28b-4637-bcce-601596e82360" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.836398] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "c0e830f3-e28b-4637-bcce-601596e82360" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.026686] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230735, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.125150] env[70013]: DEBUG nova.scheduler.client.report [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 774.160254] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279d883-6aae-cd9e-4730-12790f47f3f6/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 774.161566] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6516d8-3ad5-4301-945a-e27adfad1ab4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.169450] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279d883-6aae-cd9e-4730-12790f47f3f6/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 774.169614] env[70013]: ERROR oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279d883-6aae-cd9e-4730-12790f47f3f6/disk-0.vmdk due to incomplete transfer. [ 774.169746] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1299f0e9-6e7c-4a8d-83cc-abac997c9a11 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.180571] env[70013]: DEBUG oslo_vmware.rw_handles [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279d883-6aae-cd9e-4730-12790f47f3f6/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 774.180824] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Uploaded image 00d5405f-9c08-411a-9d32-f09745461ee9 to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 774.186715] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 774.187440] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8782e9ed-f101-46eb-ac02-6c31777b0c48 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.196107] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 774.196107] env[70013]: value = "task-4230737" [ 774.196107] env[70013]: _type = "Task" [ 774.196107] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.205928] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230737, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.519972] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230735, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555479} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.520356] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2c0e4326-c33a-42bb-b793-a100157b1c03/2c0e4326-c33a-42bb-b793-a100157b1c03.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.520627] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.520955] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5735ddfb-714b-4dba-8331-28fcd1edbc89 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.529425] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 774.529425] env[70013]: value = "task-4230738" [ 774.529425] env[70013]: _type = "Task" [ 774.529425] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.540709] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230738, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.632270] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.246s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.632832] env[70013]: DEBUG nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 774.636055] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.052s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.638393] env[70013]: INFO nova.compute.claims [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.708067] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230737, 'name': Destroy_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.044316] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230738, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.273268} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.044316] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.044316] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de02e30f-05b1-4037-bc0f-8be822280234 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.076244] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 2c0e4326-c33a-42bb-b793-a100157b1c03/2c0e4326-c33a-42bb-b793-a100157b1c03.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.076244] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70a920a4-931e-43ee-8c1f-baeb834287f4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.098110] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 775.098110] env[70013]: value = "task-4230739" [ 775.098110] env[70013]: _type = "Task" [ 775.098110] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.107707] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.143738] env[70013]: DEBUG nova.compute.utils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 775.147486] env[70013]: DEBUG nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 775.147712] env[70013]: DEBUG nova.network.neutron [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 775.209933] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230737, 'name': Destroy_Task, 'duration_secs': 0.796266} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.210235] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Destroyed the VM [ 775.210536] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 775.210701] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-055dba3c-826f-4314-8653-8db62efa30c7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.219663] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 775.219663] env[70013]: value = "task-4230740" [ 775.219663] env[70013]: _type = "Task" [ 775.219663] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.228594] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230740, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.235221] env[70013]: DEBUG nova.policy [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a7386d27973488ea728d9f442f3b47e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbdf4732a346424b8ff1c39c7e900f08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 775.527481] env[70013]: DEBUG nova.network.neutron [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Successfully created port: 49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.614195] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.616089] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "f715cf45-81f1-4979-abfc-455b3e64f00e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.616440] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.654024] env[70013]: DEBUG nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 775.732025] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230740, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.894852] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 775.894852] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837081', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'name': 'volume-d827861d-a75d-44ae-8fdc-a94870d66776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ec4b86d-e83b-4b28-ac4b-d990177b0ba5', 'attached_at': '', 'detached_at': '', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'serial': 'd827861d-a75d-44ae-8fdc-a94870d66776'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 775.894852] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209cdb0d-15c2-4346-b863-d581ec3adb87 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.914488] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5853e9-4376-490f-95c2-8edb206e181f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.943338] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] volume-d827861d-a75d-44ae-8fdc-a94870d66776/volume-d827861d-a75d-44ae-8fdc-a94870d66776.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.946879] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c914ac07-0f62-4de2-b032-c8c4c22eb2bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.967580] env[70013]: DEBUG oslo_vmware.api [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 775.967580] env[70013]: value = "task-4230741" [ 775.967580] env[70013]: _type = "Task" [ 775.967580] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.979792] env[70013]: DEBUG oslo_vmware.api [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230741, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.115114] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.169904] env[70013]: DEBUG nova.compute.manager [None req-8da3d572-7c58-4117-be25-760082593c71 tempest-ServerDiagnosticsV248Test-1970149035 tempest-ServerDiagnosticsV248Test-1970149035-project-admin] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 776.173711] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd202457-af3d-4491-a65f-4f28d1d48b94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.182122] env[70013]: INFO nova.compute.manager [None req-8da3d572-7c58-4117-be25-760082593c71 tempest-ServerDiagnosticsV248Test-1970149035 tempest-ServerDiagnosticsV248Test-1970149035-project-admin] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Retrieving diagnostics [ 776.183046] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec82dcf-7574-4baa-955b-e74ef62bf28d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.231408] env[70013]: DEBUG oslo_vmware.api [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230740, 'name': RemoveSnapshot_Task, 'duration_secs': 0.741571} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.235022] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 776.235022] env[70013]: INFO nova.compute.manager [None req-a0c99dfd-9022-4e13-b0b6-d0a07a67c32f tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Took 16.53 seconds to snapshot the instance on the hypervisor. [ 776.247260] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b366034e-4961-4405-a556-df523ce0a659 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.255534] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0812c3d4-71a4-4f3a-a730-3873cce9e0a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.286958] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c552cff-d519-4cd9-9978-0138fe4b257b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.296463] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d4efc6-52bd-4a2d-94e9-f22a511f8ef6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.315134] env[70013]: DEBUG nova.compute.provider_tree [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.478814] env[70013]: DEBUG oslo_vmware.api [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230741, 'name': ReconfigVM_Task, 'duration_secs': 0.401276} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.479158] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Reconfigured VM instance instance-0000000d to attach disk [datastore1] volume-d827861d-a75d-44ae-8fdc-a94870d66776/volume-d827861d-a75d-44ae-8fdc-a94870d66776.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.484169] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc5e6f6d-2482-4cbe-b0f1-685effbcede8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.502372] env[70013]: DEBUG oslo_vmware.api [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 776.502372] env[70013]: value = "task-4230742" [ 776.502372] env[70013]: _type = "Task" [ 776.502372] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.511616] env[70013]: DEBUG oslo_vmware.api [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.612218] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230739, 'name': ReconfigVM_Task, 'duration_secs': 1.482682} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.612572] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 2c0e4326-c33a-42bb-b793-a100157b1c03/2c0e4326-c33a-42bb-b793-a100157b1c03.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.613357] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38048d92-f594-4056-9da3-c773c588998e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.621769] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 776.621769] env[70013]: value = "task-4230743" [ 776.621769] env[70013]: _type = "Task" [ 776.621769] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.633599] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230743, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.675608] env[70013]: DEBUG nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 776.704818] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 776.704818] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.704818] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 776.705053] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.705053] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 776.705053] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 776.706440] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 776.706639] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 776.706876] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 776.707112] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 776.707347] env[70013]: DEBUG nova.virt.hardware [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 776.708696] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea01b7cf-610d-499b-8423-0cf6a1529c8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.717695] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b463946a-8387-42f9-9188-41b3732636a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.819487] env[70013]: DEBUG nova.scheduler.client.report [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 777.013752] env[70013]: DEBUG oslo_vmware.api [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230742, 'name': ReconfigVM_Task, 'duration_secs': 0.178222} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.016342] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837081', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'name': 'volume-d827861d-a75d-44ae-8fdc-a94870d66776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ec4b86d-e83b-4b28-ac4b-d990177b0ba5', 'attached_at': '', 'detached_at': '', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'serial': 'd827861d-a75d-44ae-8fdc-a94870d66776'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 777.134374] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230743, 'name': Rename_Task, 'duration_secs': 0.279419} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.134676] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.134926] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5f7beba-664a-4cbd-8c66-c93b4f61986f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.144022] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 777.144022] env[70013]: value = "task-4230744" [ 777.144022] env[70013]: _type = "Task" [ 777.144022] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.155926] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.325778] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.690s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.326668] env[70013]: DEBUG nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 777.329691] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.093s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.334468] env[70013]: INFO nova.compute.claims [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.368581] env[70013]: DEBUG nova.compute.manager [req-f95d123f-7e00-4543-99fd-e099ee9dc668 req-9857afcc-a1f2-4adc-9623-f44fefbe5b92 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Received event network-vif-plugged-49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 777.368581] env[70013]: DEBUG oslo_concurrency.lockutils [req-f95d123f-7e00-4543-99fd-e099ee9dc668 req-9857afcc-a1f2-4adc-9623-f44fefbe5b92 service nova] Acquiring lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.368581] env[70013]: DEBUG oslo_concurrency.lockutils [req-f95d123f-7e00-4543-99fd-e099ee9dc668 req-9857afcc-a1f2-4adc-9623-f44fefbe5b92 service nova] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.369981] env[70013]: DEBUG oslo_concurrency.lockutils [req-f95d123f-7e00-4543-99fd-e099ee9dc668 req-9857afcc-a1f2-4adc-9623-f44fefbe5b92 service nova] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.369981] env[70013]: DEBUG nova.compute.manager [req-f95d123f-7e00-4543-99fd-e099ee9dc668 req-9857afcc-a1f2-4adc-9623-f44fefbe5b92 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] No waiting events found dispatching network-vif-plugged-49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 777.369981] env[70013]: WARNING nova.compute.manager [req-f95d123f-7e00-4543-99fd-e099ee9dc668 req-9857afcc-a1f2-4adc-9623-f44fefbe5b92 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Received unexpected event network-vif-plugged-49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd for instance with vm_state building and task_state spawning. [ 777.373249] env[70013]: DEBUG nova.network.neutron [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Successfully updated port: 49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.465560] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "062eff58-95fe-4c9b-a586-7e7434c77adf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.465907] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "062eff58-95fe-4c9b-a586-7e7434c77adf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.466269] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "062eff58-95fe-4c9b-a586-7e7434c77adf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.466585] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "062eff58-95fe-4c9b-a586-7e7434c77adf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.467061] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "062eff58-95fe-4c9b-a586-7e7434c77adf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.470279] env[70013]: INFO nova.compute.manager [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Terminating instance [ 777.655125] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230744, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.845867] env[70013]: DEBUG nova.compute.utils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 777.847830] env[70013]: DEBUG nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 777.848709] env[70013]: DEBUG nova.network.neutron [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.875139] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.875139] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.875139] env[70013]: DEBUG nova.network.neutron [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.920238] env[70013]: DEBUG nova.policy [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '99104b73f6174c05a5272f023612bb49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f85175ef2b004d5ea372f2dfe69549ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 777.975660] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "refresh_cache-062eff58-95fe-4c9b-a586-7e7434c77adf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.975868] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquired lock "refresh_cache-062eff58-95fe-4c9b-a586-7e7434c77adf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.976329] env[70013]: DEBUG nova.network.neutron [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.077082] env[70013]: DEBUG nova.objects.instance [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lazy-loading 'flavor' on Instance uuid 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.153868] env[70013]: DEBUG oslo_vmware.api [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230744, 'name': PowerOnVM_Task, 'duration_secs': 0.724872} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.154177] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.154388] env[70013]: INFO nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Took 11.39 seconds to spawn the instance on the hypervisor. [ 778.154567] env[70013]: DEBUG nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 778.155414] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615fdd21-9b46-49cf-a4e4-dac5b4ca8294 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.284911] env[70013]: DEBUG nova.network.neutron [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Successfully created port: 9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.352260] env[70013]: DEBUG nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 778.467900] env[70013]: DEBUG nova.network.neutron [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.502990] env[70013]: DEBUG nova.network.neutron [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.587213] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7123164-f516-464d-80b1-a7bbeaa707ff tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.383s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.617265] env[70013]: DEBUG nova.network.neutron [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.675649] env[70013]: INFO nova.compute.manager [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Took 43.08 seconds to build instance. [ 778.709474] env[70013]: DEBUG nova.network.neutron [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Updating instance_info_cache with network_info: [{"id": "49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd", "address": "fa:16:3e:9f:9d:31", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49f610ba-b8", "ovs_interfaceid": "49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.905929] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd6a398-4b7a-4444-b8a3-da482e761c61 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.914792] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb47851-24e4-4929-99e1-e2ca889a3ab1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.945228] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d93691-c0b4-4e33-bd17-c43449f4b888 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.954154] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a02f5a-b626-41ce-aff9-b204c9c69f78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.970182] env[70013]: DEBUG nova.compute.provider_tree [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.123053] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Releasing lock "refresh_cache-062eff58-95fe-4c9b-a586-7e7434c77adf" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.123053] env[70013]: DEBUG nova.compute.manager [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 779.123053] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 779.123053] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc130dbd-5dd8-4506-b6df-ee3b4f73a821 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.135062] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 779.135345] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd8d6233-e4f9-44f7-b55a-6410c7b2f49b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.142940] env[70013]: DEBUG oslo_vmware.api [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 779.142940] env[70013]: value = "task-4230745" [ 779.142940] env[70013]: _type = "Task" [ 779.142940] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.154708] env[70013]: DEBUG oslo_vmware.api [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230745, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.178628] env[70013]: DEBUG oslo_concurrency.lockutils [None req-940e4ffd-d7d9-4f86-becc-fc9648184986 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.041s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.212182] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Releasing lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.212562] env[70013]: DEBUG nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance network_info: |[{"id": "49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd", "address": "fa:16:3e:9f:9d:31", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49f610ba-b8", "ovs_interfaceid": "49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 779.213214] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:9d:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.221555] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 779.221777] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 779.222087] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b28d6a9-207f-4cc5-8c6a-2752c5f4c00a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.244268] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.244268] env[70013]: value = "task-4230746" [ 779.244268] env[70013]: _type = "Task" [ 779.244268] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.252456] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230746, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.362077] env[70013]: DEBUG nova.compute.manager [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 779.363073] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3617e87c-297e-49b8-badd-55a4d74e8fcd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.369270] env[70013]: DEBUG nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 779.410520] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 779.410845] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.411055] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 779.417308] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.417693] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 779.417901] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 779.418170] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 779.418343] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 779.418523] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 779.418695] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 779.419195] env[70013]: DEBUG nova.virt.hardware [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 779.419853] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a752a3b-526e-4c61-ac62-69730e3e2c16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.429158] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844e6ea4-bf24-436c-bf75-3ba6b2ed099d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.473265] env[70013]: DEBUG nova.scheduler.client.report [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 779.657242] env[70013]: DEBUG oslo_vmware.api [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230745, 'name': PowerOffVM_Task, 'duration_secs': 0.276548} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.657519] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 779.657681] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 779.657924] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-791eb970-04ae-443a-993f-efb413fd6f3d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.681985] env[70013]: DEBUG nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 779.691017] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 779.691322] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 779.691520] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Deleting the datastore file [datastore2] 062eff58-95fe-4c9b-a586-7e7434c77adf {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.691790] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9f85a17-06e8-4c02-87c0-13be1472314d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.705495] env[70013]: DEBUG oslo_vmware.api [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for the task: (returnval){ [ 779.705495] env[70013]: value = "task-4230748" [ 779.705495] env[70013]: _type = "Task" [ 779.705495] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.716946] env[70013]: DEBUG oslo_vmware.api [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.756946] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230746, 'name': CreateVM_Task, 'duration_secs': 0.36673} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.757160] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.757851] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.758013] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.758328] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 779.758610] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbe5e1eb-b29d-4fb9-91df-35a3e8114fba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.764259] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 779.764259] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52be1caf-a8b8-2c7c-51a5-629be1cbb99e" [ 779.764259] env[70013]: _type = "Task" [ 779.764259] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.775268] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52be1caf-a8b8-2c7c-51a5-629be1cbb99e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.840472] env[70013]: DEBUG nova.compute.manager [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Received event network-changed-49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 779.840661] env[70013]: DEBUG nova.compute.manager [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Refreshing instance network info cache due to event network-changed-49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 779.840873] env[70013]: DEBUG oslo_concurrency.lockutils [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] Acquiring lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.841018] env[70013]: DEBUG oslo_concurrency.lockutils [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] Acquired lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.841176] env[70013]: DEBUG nova.network.neutron [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Refreshing network info cache for port 49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.852394] env[70013]: DEBUG nova.compute.manager [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Received event network-changed-aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 779.852502] env[70013]: DEBUG nova.compute.manager [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Refreshing instance network info cache due to event network-changed-aab77ffc-40a3-47ea-abe4-1abb8298eba0. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 779.852693] env[70013]: DEBUG oslo_concurrency.lockutils [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] Acquiring lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.852907] env[70013]: DEBUG oslo_concurrency.lockutils [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] Acquired lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.854176] env[70013]: DEBUG nova.network.neutron [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Refreshing network info cache for port aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.882503] env[70013]: INFO nova.compute.manager [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] instance snapshotting [ 779.887061] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dadb5ca1-249c-4cdb-930c-981c0a524345 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.905069] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867ae482-adec-4388-92a5-0be7110a183f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.979513] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.650s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.980108] env[70013]: DEBUG nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 779.987908] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.987908] env[70013]: DEBUG nova.objects.instance [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lazy-loading 'resources' on Instance uuid fc0b7f87-8a36-440e-b158-30b9d9920572 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 780.153926] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.153926] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.203566] env[70013]: DEBUG nova.network.neutron [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Successfully updated port: 9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.216699] env[70013]: DEBUG oslo_vmware.api [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Task: {'id': task-4230748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101725} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.216699] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.217103] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 780.217103] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.217103] env[70013]: INFO nova.compute.manager [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Took 1.10 seconds to destroy the instance on the hypervisor. [ 780.217103] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 780.217989] env[70013]: DEBUG nova.compute.manager [-] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 780.218060] env[70013]: DEBUG nova.network.neutron [-] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.233230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.247925] env[70013]: DEBUG nova.network.neutron [-] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.281384] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52be1caf-a8b8-2c7c-51a5-629be1cbb99e, 'name': SearchDatastore_Task, 'duration_secs': 0.010053} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.281719] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 780.281960] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.282215] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.282364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 780.282540] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.282802] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff269bd2-9fc4-42ae-ae3e-e645c66a7703 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.292647] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.292842] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 780.293727] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29edf5f4-f0fc-4e10-a13c-7d7704dcdff5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.300297] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 780.300297] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab017f-0de1-ef9f-23ce-83d37bfcd76d" [ 780.300297] env[70013]: _type = "Task" [ 780.300297] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.309855] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab017f-0de1-ef9f-23ce-83d37bfcd76d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.421616] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 780.421814] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-08d20f6e-23f8-4ec2-89ab-ef79e7bf5676 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.435137] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 780.435137] env[70013]: value = "task-4230749" [ 780.435137] env[70013]: _type = "Task" [ 780.435137] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.448600] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230749, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.492464] env[70013]: DEBUG nova.compute.utils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 780.497486] env[70013]: DEBUG nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 780.498762] env[70013]: DEBUG nova.network.neutron [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 780.642174] env[70013]: DEBUG nova.policy [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6588ea7b79c413cba9eee31d01dd3f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '02168307ea9347dd8f6d6993dc1f9b2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 780.656190] env[70013]: INFO nova.compute.manager [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Detaching volume d827861d-a75d-44ae-8fdc-a94870d66776 [ 780.707425] env[70013]: INFO nova.virt.block_device [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Attempting to driver detach volume d827861d-a75d-44ae-8fdc-a94870d66776 from mountpoint /dev/sdb [ 780.707762] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 780.708116] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837081', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'name': 'volume-d827861d-a75d-44ae-8fdc-a94870d66776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ec4b86d-e83b-4b28-ac4b-d990177b0ba5', 'attached_at': '', 'detached_at': '', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'serial': 'd827861d-a75d-44ae-8fdc-a94870d66776'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 780.709359] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0602de-543d-4415-8925-b82eec8b8532 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.717530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.717530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 780.717868] env[70013]: DEBUG nova.network.neutron [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.753820] env[70013]: DEBUG nova.network.neutron [-] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.758418] env[70013]: DEBUG nova.network.neutron [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Updated VIF entry in instance network info cache for port 49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 780.758768] env[70013]: DEBUG nova.network.neutron [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Updating instance_info_cache with network_info: [{"id": "49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd", "address": "fa:16:3e:9f:9d:31", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49f610ba-b8", "ovs_interfaceid": "49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.760570] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170bc7e1-caf4-45c0-bfe6-a2d4b1b46882 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.771107] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37453327-4b07-45a8-bfdb-4ddefd8d5e9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.807250] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318daa27-ec6c-4532-b409-1497be184ddf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.813165] env[70013]: DEBUG nova.network.neutron [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.832019] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] The volume has not been displaced from its original location: [datastore1] volume-d827861d-a75d-44ae-8fdc-a94870d66776/volume-d827861d-a75d-44ae-8fdc-a94870d66776.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 780.838221] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Reconfiguring VM instance instance-0000000d to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 780.846376] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4eba5da-a66c-445b-af90-f7d6b7e73506 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.860532] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab017f-0de1-ef9f-23ce-83d37bfcd76d, 'name': SearchDatastore_Task, 'duration_secs': 0.010263} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.864680] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-441fa35a-780a-434c-8809-3aaa24459b25 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.870456] env[70013]: DEBUG oslo_vmware.api [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 780.870456] env[70013]: value = "task-4230750" [ 780.870456] env[70013]: _type = "Task" [ 780.870456] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.873292] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 780.873292] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f257d3-fee3-7d65-22fa-abdd312f08ad" [ 780.873292] env[70013]: _type = "Task" [ 780.873292] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.890729] env[70013]: DEBUG oslo_vmware.api [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230750, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.895430] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f257d3-fee3-7d65-22fa-abdd312f08ad, 'name': SearchDatastore_Task, 'duration_secs': 0.014631} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.896116] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 780.896116] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 85488f5e-bd8a-4bcf-8420-744c54176c2e/85488f5e-bd8a-4bcf-8420-744c54176c2e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.896366] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44d25c24-e4ef-46f0-a2fc-744972596236 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.907659] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 780.907659] env[70013]: value = "task-4230751" [ 780.907659] env[70013]: _type = "Task" [ 780.907659] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.912714] env[70013]: DEBUG nova.network.neutron [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updated VIF entry in instance network info cache for port aab77ffc-40a3-47ea-abe4-1abb8298eba0. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 780.912714] env[70013]: DEBUG nova.network.neutron [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating instance_info_cache with network_info: [{"id": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "address": "fa:16:3e:5d:0d:2d", "network": {"id": "408751e1-e468-4c61-b5c6-06c709856e70", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-745309513-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f0e36e7872c4024b587dc70b9a15ea0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab77ffc-40", "ovs_interfaceid": "aab77ffc-40a3-47ea-abe4-1abb8298eba0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.921398] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.951187] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230749, 'name': CreateSnapshot_Task, 'duration_secs': 0.510742} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.951626] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 780.953120] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cd1070-1f70-425e-94ae-35404b653942 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.998630] env[70013]: DEBUG nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 781.093955] env[70013]: DEBUG nova.network.neutron [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updating instance_info_cache with network_info: [{"id": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "address": "fa:16:3e:93:5e:8d", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e2b8cbf-ea", "ovs_interfaceid": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.171052] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ff4aec-7777-4048-82a9-1d6c73242fbc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.179875] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f073ed70-e6d8-468b-b058-aa7d7c80cb90 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.211709] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0245e586-0742-455c-a32c-d0ce65eeed07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.220354] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916c8a79-7429-42ce-b016-2886557e1d56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.237100] env[70013]: DEBUG nova.compute.provider_tree [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.265128] env[70013]: DEBUG oslo_concurrency.lockutils [req-314a48b1-2094-4121-890b-6987d2ea68ea req-6f0c0beb-9cc7-4d35-9e94-53da04535510 service nova] Releasing lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.265700] env[70013]: INFO nova.compute.manager [-] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Took 1.05 seconds to deallocate network for instance. [ 781.384048] env[70013]: DEBUG oslo_vmware.api [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230750, 'name': ReconfigVM_Task, 'duration_secs': 0.277862} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.384481] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Reconfigured VM instance instance-0000000d to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 781.390292] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d99b926-bf20-429e-8d96-77628cf4dbc7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.411873] env[70013]: DEBUG oslo_vmware.api [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 781.411873] env[70013]: value = "task-4230752" [ 781.411873] env[70013]: _type = "Task" [ 781.411873] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.419242] env[70013]: DEBUG oslo_concurrency.lockutils [req-26671dca-4702-4249-89e2-80c1419467c7 req-1550d53b-edc0-41da-9c72-4e5fe52966a6 service nova] Releasing lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.435696] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230751, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.443397] env[70013]: DEBUG oslo_vmware.api [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.477491] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 781.477772] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b99c855c-9ee0-4de9-96dc-d9b784b170b4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.487705] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 781.487705] env[70013]: value = "task-4230753" [ 781.487705] env[70013]: _type = "Task" [ 781.487705] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.499358] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230753, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.596632] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Releasing lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.597111] env[70013]: DEBUG nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Instance network_info: |[{"id": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "address": "fa:16:3e:93:5e:8d", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e2b8cbf-ea", "ovs_interfaceid": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 781.597682] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:5e:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e2b8cbf-ea68-4961-ae17-64f6753e3974', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.606184] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Creating folder: Project (f85175ef2b004d5ea372f2dfe69549ec). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.607169] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-999d477a-1182-48e6-a8cc-3fabe8835175 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.621794] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Created folder: Project (f85175ef2b004d5ea372f2dfe69549ec) in parent group-v836999. [ 781.621794] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Creating folder: Instances. Parent ref: group-v837085. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.621794] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce764ee3-938d-4a1f-a133-6f58ff26833b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.636425] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Created folder: Instances in parent group-v837085. [ 781.636702] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 781.636940] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 781.637407] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7a4c3df-6bb9-474f-93b0-95479a23ecd0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.653438] env[70013]: DEBUG nova.network.neutron [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Successfully created port: 42a0109b-8ff8-4977-b5ce-f0caa32df87e {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.661904] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.661904] env[70013]: value = "task-4230756" [ 781.661904] env[70013]: _type = "Task" [ 781.661904] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.671926] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230756, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.741695] env[70013]: DEBUG nova.scheduler.client.report [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 781.780981] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 781.935024] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.72352} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.935024] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 85488f5e-bd8a-4bcf-8420-744c54176c2e/85488f5e-bd8a-4bcf-8420-744c54176c2e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 781.935024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.935024] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5d7d574-712b-4734-a181-56da356b3ea1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.941243] env[70013]: DEBUG oslo_vmware.api [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230752, 'name': ReconfigVM_Task, 'duration_secs': 0.202333} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.942184] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837081', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'name': 'volume-d827861d-a75d-44ae-8fdc-a94870d66776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ec4b86d-e83b-4b28-ac4b-d990177b0ba5', 'attached_at': '', 'detached_at': '', 'volume_id': 'd827861d-a75d-44ae-8fdc-a94870d66776', 'serial': 'd827861d-a75d-44ae-8fdc-a94870d66776'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 781.950999] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 781.950999] env[70013]: value = "task-4230757" [ 781.950999] env[70013]: _type = "Task" [ 781.950999] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.963332] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230757, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.001489] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230753, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.013120] env[70013]: DEBUG nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 782.054254] env[70013]: DEBUG nova.compute.manager [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Received event network-vif-plugged-9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 782.056627] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] Acquiring lock "67b90fec-a6fd-471d-a425-82bc63de572a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 782.060028] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] Lock "67b90fec-a6fd-471d-a425-82bc63de572a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 782.060028] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] Lock "67b90fec-a6fd-471d-a425-82bc63de572a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.060028] env[70013]: DEBUG nova.compute.manager [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] No waiting events found dispatching network-vif-plugged-9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 782.060028] env[70013]: WARNING nova.compute.manager [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Received unexpected event network-vif-plugged-9e2b8cbf-ea68-4961-ae17-64f6753e3974 for instance with vm_state building and task_state spawning. [ 782.060028] env[70013]: DEBUG nova.compute.manager [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Received event network-changed-9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 782.060461] env[70013]: DEBUG nova.compute.manager [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Refreshing instance network info cache due to event network-changed-9e2b8cbf-ea68-4961-ae17-64f6753e3974. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 782.060461] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] Acquiring lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.060461] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] Acquired lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.060461] env[70013]: DEBUG nova.network.neutron [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Refreshing network info cache for port 9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.063535] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 782.063742] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.063907] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 782.064104] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.064293] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 782.064398] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 782.064603] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 782.064761] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 782.064928] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 782.065104] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 782.065354] env[70013]: DEBUG nova.virt.hardware [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 782.068166] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec1afb2-4504-4cbf-bd92-d313e62046fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.080038] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8b1d5b-5939-4aa6-bf70-1b8af97b78ed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.178160] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230756, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.248244] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.261s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.251706] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.906s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 782.251893] env[70013]: DEBUG nova.objects.instance [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'resources' on Instance uuid fc473d40-b57b-437e-9511-58a0ba700a69 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 782.273347] env[70013]: INFO nova.scheduler.client.report [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Deleted allocations for instance fc0b7f87-8a36-440e-b158-30b9d9920572 [ 782.462861] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230757, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.180143} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.463398] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 782.464224] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bedb5df-5b18-4a15-9569-56af96332020 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.489187] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] 85488f5e-bd8a-4bcf-8420-744c54176c2e/85488f5e-bd8a-4bcf-8420-744c54176c2e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 782.489655] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c095d8e-1859-47a5-aa27-2de7784caa40 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.514195] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230753, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.515724] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 782.515724] env[70013]: value = "task-4230758" [ 782.515724] env[70013]: _type = "Task" [ 782.515724] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.522310] env[70013]: DEBUG nova.objects.instance [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lazy-loading 'flavor' on Instance uuid 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 782.675259] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230756, 'name': CreateVM_Task, 'duration_secs': 0.811836} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.675494] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.676285] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.676453] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.676788] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 782.677498] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1319f126-774c-4658-8ddb-a25045dc80a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.683243] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 782.683243] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d6578d-ce74-ffa7-3333-2308347505bb" [ 782.683243] env[70013]: _type = "Task" [ 782.683243] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.693257] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d6578d-ce74-ffa7-3333-2308347505bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.793577] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ac494991-daa2-4135-b225-7e41193ada96 tempest-ServerExternalEventsTest-1174324192 tempest-ServerExternalEventsTest-1174324192-project-member] Lock "fc0b7f87-8a36-440e-b158-30b9d9920572" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.445s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.004954] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230753, 'name': CloneVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.017988] env[70013]: DEBUG nova.network.neutron [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updated VIF entry in instance network info cache for port 9e2b8cbf-ea68-4961-ae17-64f6753e3974. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 783.018392] env[70013]: DEBUG nova.network.neutron [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updating instance_info_cache with network_info: [{"id": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "address": "fa:16:3e:93:5e:8d", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e2b8cbf-ea", "ovs_interfaceid": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.037720] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230758, 'name': ReconfigVM_Task, 'duration_secs': 0.486469} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.039069] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Reconfigured VM instance instance-00000018 to attach disk [datastore1] 85488f5e-bd8a-4bcf-8420-744c54176c2e/85488f5e-bd8a-4bcf-8420-744c54176c2e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.039862] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14ed8ee6-6670-48a8-8615-e73096b1e7f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.047752] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 783.047752] env[70013]: value = "task-4230759" [ 783.047752] env[70013]: _type = "Task" [ 783.047752] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.060250] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230759, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.201080] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d6578d-ce74-ffa7-3333-2308347505bb, 'name': SearchDatastore_Task, 'duration_secs': 0.010161} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.201678] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.202035] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.202382] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.202625] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.202920] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.203288] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81a1f344-edd6-4402-8aee-bf39d5b6f565 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.220714] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.221923] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.222635] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e95e0c0e-8aa7-4e60-af57-1fbb575ff6a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.232885] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 783.232885] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52599b4d-18ae-a1d3-1478-8476864a0115" [ 783.232885] env[70013]: _type = "Task" [ 783.232885] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.243066] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52599b4d-18ae-a1d3-1478-8476864a0115, 'name': SearchDatastore_Task, 'duration_secs': 0.010069} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.244108] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-776cf857-4b47-4264-860c-a9c81e955733 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.255090] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 783.255090] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5203ca36-ea34-b119-5053-094ea65b7e88" [ 783.255090] env[70013]: _type = "Task" [ 783.255090] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.264698] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5203ca36-ea34-b119-5053-094ea65b7e88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.280673] env[70013]: DEBUG nova.network.neutron [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Successfully updated port: 42a0109b-8ff8-4977-b5ce-f0caa32df87e {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.301656] env[70013]: DEBUG nova.compute.manager [req-c9d92a44-b1a6-4a42-941b-c8ef3f78b981 req-f66e22e9-d883-41f0-8aad-30282c7a1af3 service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Received event network-vif-plugged-42a0109b-8ff8-4977-b5ce-f0caa32df87e {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 783.301887] env[70013]: DEBUG oslo_concurrency.lockutils [req-c9d92a44-b1a6-4a42-941b-c8ef3f78b981 req-f66e22e9-d883-41f0-8aad-30282c7a1af3 service nova] Acquiring lock "cfe1478c-53f2-4df8-99a8-ab23209f5306-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.302123] env[70013]: DEBUG oslo_concurrency.lockutils [req-c9d92a44-b1a6-4a42-941b-c8ef3f78b981 req-f66e22e9-d883-41f0-8aad-30282c7a1af3 service nova] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.302295] env[70013]: DEBUG oslo_concurrency.lockutils [req-c9d92a44-b1a6-4a42-941b-c8ef3f78b981 req-f66e22e9-d883-41f0-8aad-30282c7a1af3 service nova] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.302460] env[70013]: DEBUG nova.compute.manager [req-c9d92a44-b1a6-4a42-941b-c8ef3f78b981 req-f66e22e9-d883-41f0-8aad-30282c7a1af3 service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] No waiting events found dispatching network-vif-plugged-42a0109b-8ff8-4977-b5ce-f0caa32df87e {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 783.302626] env[70013]: WARNING nova.compute.manager [req-c9d92a44-b1a6-4a42-941b-c8ef3f78b981 req-f66e22e9-d883-41f0-8aad-30282c7a1af3 service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Received unexpected event network-vif-plugged-42a0109b-8ff8-4977-b5ce-f0caa32df87e for instance with vm_state building and task_state spawning. [ 783.377467] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaa0144-c242-4032-9fe3-645b5b51d450 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.386648] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78102236-aba2-4d67-815c-314ef808b619 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.434588] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cdad84-5ada-4b65-8606-b0eb54a1dd51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.443708] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ed0ac6-3e57-4f3e-92cd-e251438b19df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.458278] env[70013]: DEBUG nova.compute.provider_tree [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.511110] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230753, 'name': CloneVM_Task, 'duration_secs': 1.555791} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.511431] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Created linked-clone VM from snapshot [ 783.513544] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157dadfd-cf2d-45a1-95dc-5bc8b1cafcff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.522695] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Uploading image d4b5d8a9-aa58-4102-b923-43a5fc4d950f {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 783.529042] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef5a1552-69d8-410c-a370-82fef53fcc9e req-163c1700-3e58-4413-ba6a-9e6be2b59eaf service nova] Releasing lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.537795] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f1c50b65-8cb3-4e44-b482-2dede271a2bc tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.386s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.563664] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 783.563664] env[70013]: value = "vm-837084" [ 783.563664] env[70013]: _type = "VirtualMachine" [ 783.563664] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 783.564018] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230759, 'name': Rename_Task, 'duration_secs': 0.150359} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.564215] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cf4ee926-5f48-4088-9a66-1a67095e7951 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.569999] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 783.569999] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24dfd5f4-c596-4338-a30b-0e17212eb705 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.576763] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lease: (returnval){ [ 783.576763] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5b60-a5dc-504b-e2b1-15c444e81928" [ 783.576763] env[70013]: _type = "HttpNfcLease" [ 783.576763] env[70013]: } obtained for exporting VM: (result){ [ 783.576763] env[70013]: value = "vm-837084" [ 783.576763] env[70013]: _type = "VirtualMachine" [ 783.576763] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 783.577071] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the lease: (returnval){ [ 783.577071] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5b60-a5dc-504b-e2b1-15c444e81928" [ 783.577071] env[70013]: _type = "HttpNfcLease" [ 783.577071] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 783.580340] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 783.580340] env[70013]: value = "task-4230761" [ 783.580340] env[70013]: _type = "Task" [ 783.580340] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.596359] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230761, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.598034] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 783.598034] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5b60-a5dc-504b-e2b1-15c444e81928" [ 783.598034] env[70013]: _type = "HttpNfcLease" [ 783.598034] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 783.598741] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 783.598741] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5b60-a5dc-504b-e2b1-15c444e81928" [ 783.598741] env[70013]: _type = "HttpNfcLease" [ 783.598741] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 783.599307] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f828dcb-9e62-466d-9606-3a44b6abc5f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.609153] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f7d5b-9053-4aea-674f-bce41eb4659a/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 783.609273] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f7d5b-9053-4aea-674f-bce41eb4659a/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 783.713195] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fcc5a327-ae09-423f-b9c6-b740bce4b046 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.767509] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5203ca36-ea34-b119-5053-094ea65b7e88, 'name': SearchDatastore_Task, 'duration_secs': 0.009777} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.767811] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.768085] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 67b90fec-a6fd-471d-a425-82bc63de572a/67b90fec-a6fd-471d-a425-82bc63de572a.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 783.768536] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4225f27d-bbad-42b4-a1d7-95fe001bbde6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.778680] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 783.778680] env[70013]: value = "task-4230762" [ 783.778680] env[70013]: _type = "Task" [ 783.778680] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.783698] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "refresh_cache-cfe1478c-53f2-4df8-99a8-ab23209f5306" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.783851] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquired lock "refresh_cache-cfe1478c-53f2-4df8-99a8-ab23209f5306" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.784107] env[70013]: DEBUG nova.network.neutron [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.791335] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230762, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.996124] env[70013]: ERROR nova.scheduler.client.report [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [req-16205499-7cc0-4a6e-a789-bd1751daf972] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-16205499-7cc0-4a6e-a789-bd1751daf972"}]} [ 783.996124] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.997750] env[70013]: ERROR nova.compute.manager [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 783.997750] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Traceback (most recent call last): [ 783.997750] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 783.997750] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] yield [ 783.997750] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 783.997750] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self.set_inventory_for_provider( [ 783.997750] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 783.997750] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-16205499-7cc0-4a6e-a789-bd1751daf972"}]} [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] During handling of the above exception, another exception occurred: [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Traceback (most recent call last): [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self._delete_instance(context, instance, bdms) [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 783.998040] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self._complete_deletion(context, instance) [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self._update_resource_tracker(context, instance) [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self.rt.update_usage(context, instance, instance.node) [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] return f(*args, **kwargs) [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self._update(context.elevated(), self.compute_nodes[nodename]) [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self._update_to_placement(context, compute_node, startup) [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 783.998266] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] return attempt.get(self._wrap_exception) [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] six.reraise(self.value[0], self.value[1], self.value[2]) [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] raise value [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self.reportclient.update_from_provider_tree( [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] with catch_all(pd.uuid): [ 783.998644] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 783.999037] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] self.gen.throw(typ, value, traceback) [ 783.999037] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 783.999037] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] raise exception.ResourceProviderSyncFailed() [ 783.999037] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 783.999037] env[70013]: ERROR nova.compute.manager [instance: fc473d40-b57b-437e-9511-58a0ba700a69] [ 784.003808] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.795s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.005643] env[70013]: INFO nova.compute.claims [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.097772] env[70013]: DEBUG oslo_vmware.api [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230761, 'name': PowerOnVM_Task, 'duration_secs': 0.514906} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.098081] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 784.098295] env[70013]: INFO nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Took 7.42 seconds to spawn the instance on the hypervisor. [ 784.098592] env[70013]: DEBUG nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 784.099609] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28030dd3-bbc8-4846-a242-6c638ef9a799 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.294975] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230762, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.362020] env[70013]: DEBUG nova.network.neutron [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.514344] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.833s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 784.624426] env[70013]: INFO nova.compute.manager [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Took 41.75 seconds to build instance. [ 784.628368] env[70013]: DEBUG nova.network.neutron [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Updating instance_info_cache with network_info: [{"id": "42a0109b-8ff8-4977-b5ce-f0caa32df87e", "address": "fa:16:3e:68:da:78", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42a0109b-8f", "ovs_interfaceid": "42a0109b-8ff8-4977-b5ce-f0caa32df87e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.799625] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230762, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519092} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.800230] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 67b90fec-a6fd-471d-a425-82bc63de572a/67b90fec-a6fd-471d-a425-82bc63de572a.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 784.800576] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.800906] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b451dc3-8f5a-4d4e-bd4c-3cda62c4b41d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.810395] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 784.810395] env[70013]: value = "task-4230763" [ 784.810395] env[70013]: _type = "Task" [ 784.810395] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.834519] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.017033] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "be9a565b-c74c-442e-8087-cc4579b1c8e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.017405] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.057128] env[70013]: DEBUG nova.scheduler.client.report [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 785.087506] env[70013]: DEBUG nova.scheduler.client.report [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 785.087894] env[70013]: DEBUG nova.compute.provider_tree [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 785.106878] env[70013]: DEBUG nova.scheduler.client.report [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 785.127125] env[70013]: DEBUG oslo_concurrency.lockutils [None req-99477257-dfbf-4ed1-836d-3b5dc227829d tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.302s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.131872] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Releasing lock "refresh_cache-cfe1478c-53f2-4df8-99a8-ab23209f5306" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 785.132673] env[70013]: DEBUG nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Instance network_info: |[{"id": "42a0109b-8ff8-4977-b5ce-f0caa32df87e", "address": "fa:16:3e:68:da:78", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42a0109b-8f", "ovs_interfaceid": "42a0109b-8ff8-4977-b5ce-f0caa32df87e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 785.133557] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:da:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42a0109b-8ff8-4977-b5ce-f0caa32df87e', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.143106] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Creating folder: Project (02168307ea9347dd8f6d6993dc1f9b2c). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.144586] env[70013]: DEBUG nova.scheduler.client.report [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 785.149133] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6af4a97b-2115-44a9-b8a5-d117b1b11903 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.156612] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.156799] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.170041] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Created folder: Project (02168307ea9347dd8f6d6993dc1f9b2c) in parent group-v836999. [ 785.170298] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Creating folder: Instances. Parent ref: group-v837088. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.170601] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-525c4eae-8504-4372-a8f9-190c496a7b39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.186554] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Created folder: Instances in parent group-v837088. [ 785.190490] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 785.193530] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.196109] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15ad8772-ec6a-4125-9528-43ef520adb2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.224183] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.224183] env[70013]: value = "task-4230766" [ 785.224183] env[70013]: _type = "Task" [ 785.224183] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.236102] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230766, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.332051] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154593} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.332384] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.333421] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e46531e-2316-40a6-9b6a-bba448b81f1d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.369916] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] 67b90fec-a6fd-471d-a425-82bc63de572a/67b90fec-a6fd-471d-a425-82bc63de572a.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.376762] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e6da9a5-d9f9-4989-95b4-71161b126157 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.401246] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 785.401246] env[70013]: value = "task-4230767" [ 785.401246] env[70013]: _type = "Task" [ 785.401246] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.410967] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230767, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.459763] env[70013]: DEBUG nova.compute.manager [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Received event network-changed-42a0109b-8ff8-4977-b5ce-f0caa32df87e {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 785.459763] env[70013]: DEBUG nova.compute.manager [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Refreshing instance network info cache due to event network-changed-42a0109b-8ff8-4977-b5ce-f0caa32df87e. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 785.459763] env[70013]: DEBUG oslo_concurrency.lockutils [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] Acquiring lock "refresh_cache-cfe1478c-53f2-4df8-99a8-ab23209f5306" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.459763] env[70013]: DEBUG oslo_concurrency.lockutils [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] Acquired lock "refresh_cache-cfe1478c-53f2-4df8-99a8-ab23209f5306" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 785.459763] env[70013]: DEBUG nova.network.neutron [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Refreshing network info cache for port 42a0109b-8ff8-4977-b5ce-f0caa32df87e {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.632161] env[70013]: DEBUG nova.compute.manager [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 785.742026] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230766, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.757020] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 785.757410] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 785.850552] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60637dd7-2f6f-4dfa-81f3-9e2ea01e3b96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.861617] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bad728-54df-4e9b-8fee-2ba302f8acd0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.897299] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65183c5f-eadd-403a-9cc0-3f6fb790bd45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.910855] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e5d4f4-11bd-4938-9b65-1cd0cabbe040 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.921381] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230767, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.935059] env[70013]: DEBUG nova.compute.provider_tree [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.038299] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.162590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.197568] env[70013]: DEBUG nova.network.neutron [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Updated VIF entry in instance network info cache for port 42a0109b-8ff8-4977-b5ce-f0caa32df87e. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 786.198011] env[70013]: DEBUG nova.network.neutron [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Updating instance_info_cache with network_info: [{"id": "42a0109b-8ff8-4977-b5ce-f0caa32df87e", "address": "fa:16:3e:68:da:78", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42a0109b-8f", "ovs_interfaceid": "42a0109b-8ff8-4977-b5ce-f0caa32df87e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.220250] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.220530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.220761] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.221272] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.221598] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.224512] env[70013]: INFO nova.compute.manager [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Terminating instance [ 786.240031] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230766, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.268539] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.269841] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.269841] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.269969] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.270292] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.270573] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.270850] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 786.271108] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.415401] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230767, 'name': ReconfigVM_Task, 'duration_secs': 0.569181} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.415401] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Reconfigured VM instance instance-00000019 to attach disk [datastore1] 67b90fec-a6fd-471d-a425-82bc63de572a/67b90fec-a6fd-471d-a425-82bc63de572a.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.415401] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-faecee76-cc3f-46f0-a9b9-1712e18ae17b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.422936] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 786.422936] env[70013]: value = "task-4230768" [ 786.422936] env[70013]: _type = "Task" [ 786.422936] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.434048] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230768, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.462368] env[70013]: ERROR nova.scheduler.client.report [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [req-6455eebc-8eb4-47b9-a598-dc41cf542afd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6455eebc-8eb4-47b9-a598-dc41cf542afd"}]} [ 786.462524] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.466021] env[70013]: ERROR nova.compute.manager [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 786.466021] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Traceback (most recent call last): [ 786.466021] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 786.466021] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] yield [ 786.466021] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 786.466021] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] self.set_inventory_for_provider( [ 786.466021] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 786.466021] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6455eebc-8eb4-47b9-a598-dc41cf542afd"}]} [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] During handling of the above exception, another exception occurred: [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Traceback (most recent call last): [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] with self.rt.instance_claim(context, instance, node, allocs, [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 786.466407] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] return f(*args, **kwargs) [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] self._update(elevated, cn) [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] self._update_to_placement(context, compute_node, startup) [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] return attempt.get(self._wrap_exception) [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] raise value [ 786.466671] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] self.reportclient.update_from_provider_tree( [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] with catch_all(pd.uuid): [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] self.gen.throw(typ, value, traceback) [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] raise exception.ResourceProviderSyncFailed() [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 786.466989] env[70013]: ERROR nova.compute.manager [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] [ 786.467331] env[70013]: DEBUG nova.compute.utils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 786.467331] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.570s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.467803] env[70013]: INFO nova.compute.claims [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.475078] env[70013]: DEBUG nova.compute.manager [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Build of instance 8317b86e-1408-484c-822a-3b2f248c3a4d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 786.475586] env[70013]: DEBUG nova.compute.manager [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 786.475803] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Acquiring lock "refresh_cache-8317b86e-1408-484c-822a-3b2f248c3a4d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.475945] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Acquired lock "refresh_cache-8317b86e-1408-484c-822a-3b2f248c3a4d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.476169] env[70013]: DEBUG nova.network.neutron [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.702178] env[70013]: DEBUG oslo_concurrency.lockutils [req-91250b03-3c74-4824-8481-66a089a2d764 req-92c20e8b-e422-46bf-a4e6-a6d60f35608d service nova] Releasing lock "refresh_cache-cfe1478c-53f2-4df8-99a8-ab23209f5306" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.736274] env[70013]: DEBUG nova.compute.manager [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 786.736274] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.744358] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96524e0-3a10-49f9-a9c7-0f21620faae9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.751544] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230766, 'name': CreateVM_Task, 'duration_secs': 1.418871} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.754053] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.754455] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.755232] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.755456] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.755829] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 786.756158] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9783e35-b0dc-46f6-831a-723b96b0a989 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.758094] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c11c414-6a02-4dea-97d2-57a9902fc47f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.767403] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 786.767403] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b59ac-a7d7-6704-f977-cee9f548186c" [ 786.767403] env[70013]: _type = "Task" [ 786.767403] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.769621] env[70013]: DEBUG oslo_vmware.api [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 786.769621] env[70013]: value = "task-4230769" [ 786.769621] env[70013]: _type = "Task" [ 786.769621] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.779736] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.783874] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b59ac-a7d7-6704-f977-cee9f548186c, 'name': SearchDatastore_Task, 'duration_secs': 0.01201} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.790799] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.791163] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.791500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.791708] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.791960] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.793101] env[70013]: DEBUG oslo_vmware.api [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.793101] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9d476e3-cd5a-4f55-9828-dd725594b062 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.805143] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.805579] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 786.807080] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18ad4c75-6a28-412b-8b2b-2e5fa55aaa96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.816405] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 786.816405] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525a3631-493c-5cfc-5b3e-553ec5eb8659" [ 786.816405] env[70013]: _type = "Task" [ 786.816405] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.839788] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "57485f78-b97e-423c-b1f1-8648e963e5dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.839788] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "57485f78-b97e-423c-b1f1-8648e963e5dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.839788] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525a3631-493c-5cfc-5b3e-553ec5eb8659, 'name': SearchDatastore_Task, 'duration_secs': 0.012946} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.840044] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9298f16a-98d4-4866-b53e-1def85b22d42 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.847800] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 786.847800] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b7dfd8-a625-3d00-e9e6-ccf27273bef1" [ 786.847800] env[70013]: _type = "Task" [ 786.847800] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.864915] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b7dfd8-a625-3d00-e9e6-ccf27273bef1, 'name': SearchDatastore_Task, 'duration_secs': 0.011291} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.865302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.865804] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] cfe1478c-53f2-4df8-99a8-ab23209f5306/cfe1478c-53f2-4df8-99a8-ab23209f5306.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 786.865902] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b956394d-3080-4a14-a103-55dfee0bc562 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.874383] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 786.874383] env[70013]: value = "task-4230770" [ 786.874383] env[70013]: _type = "Task" [ 786.874383] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.885161] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230770, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.936301] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230768, 'name': Rename_Task, 'duration_secs': 0.156702} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.936301] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 786.936301] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07d22ad3-0423-4b33-9d26-9df31f120ee4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.944580] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 786.944580] env[70013]: value = "task-4230771" [ 786.944580] env[70013]: _type = "Task" [ 786.944580] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.956107] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230771, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.999902] env[70013]: DEBUG nova.network.neutron [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.078108] env[70013]: DEBUG nova.network.neutron [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.286022] env[70013]: DEBUG oslo_vmware.api [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230769, 'name': PowerOffVM_Task, 'duration_secs': 0.228447} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.286413] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 787.286613] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 787.286908] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11c1fd86-a04a-4645-80f8-3e7ed1331619 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.370248] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 787.370603] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 787.370821] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Deleting the datastore file [datastore1] 85488f5e-bd8a-4bcf-8420-744c54176c2e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.373181] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f1f773e-5b96-48c9-85f7-82cd1356303b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.388035] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230770, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.389565] env[70013]: DEBUG oslo_vmware.api [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for the task: (returnval){ [ 787.389565] env[70013]: value = "task-4230773" [ 787.389565] env[70013]: _type = "Task" [ 787.389565] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.400543] env[70013]: DEBUG oslo_vmware.api [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.455357] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230771, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.499654] env[70013]: DEBUG nova.scheduler.client.report [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 787.515603] env[70013]: DEBUG nova.scheduler.client.report [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 787.515832] env[70013]: DEBUG nova.compute.provider_tree [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 787.529310] env[70013]: DEBUG nova.scheduler.client.report [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 787.550009] env[70013]: DEBUG nova.scheduler.client.report [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 787.581686] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Releasing lock "refresh_cache-8317b86e-1408-484c-822a-3b2f248c3a4d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.581974] env[70013]: DEBUG nova.compute.manager [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 787.582194] env[70013]: DEBUG nova.compute.manager [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 787.582372] env[70013]: DEBUG nova.network.neutron [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.603513] env[70013]: DEBUG nova.network.neutron [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.887430] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230770, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569188} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.887705] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] cfe1478c-53f2-4df8-99a8-ab23209f5306/cfe1478c-53f2-4df8-99a8-ab23209f5306.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 787.887928] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 787.888210] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-518a5afb-4792-4c92-bab2-0df33151ab60 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.907603] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 787.907603] env[70013]: value = "task-4230774" [ 787.907603] env[70013]: _type = "Task" [ 787.907603] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.907603] env[70013]: DEBUG oslo_vmware.api [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Task: {'id': task-4230773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208058} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.907861] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.908083] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.908269] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.908447] env[70013]: INFO nova.compute.manager [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 787.908684] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 787.912252] env[70013]: DEBUG nova.compute.manager [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 787.912397] env[70013]: DEBUG nova.network.neutron [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.921532] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230774, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.959261] env[70013]: DEBUG oslo_vmware.api [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230771, 'name': PowerOnVM_Task, 'duration_secs': 0.747566} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.959261] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 787.959463] env[70013]: INFO nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Took 8.59 seconds to spawn the instance on the hypervisor. [ 787.959603] env[70013]: DEBUG nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 787.960457] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21066da-5aa4-461b-ad4e-383119bc02e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.108757] env[70013]: DEBUG nova.network.neutron [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.169504] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3175f645-1628-4c75-a3ad-15a44e889718 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.179468] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374065b7-42f2-4463-a01a-535d5cdec760 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.219703] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efb4704-c517-4046-a843-ccaef2472079 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.231200] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5377764f-1729-43c3-a38c-766dd0709e2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.249091] env[70013]: DEBUG nova.compute.provider_tree [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 788.254511] env[70013]: DEBUG nova.compute.manager [req-46ca0ee7-5d32-46a0-9ecf-e00c21530367 req-28024c91-cae9-4426-9a83-ce03f7ad2609 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Received event network-vif-deleted-49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 788.255056] env[70013]: INFO nova.compute.manager [req-46ca0ee7-5d32-46a0-9ecf-e00c21530367 req-28024c91-cae9-4426-9a83-ce03f7ad2609 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Neutron deleted interface 49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd; detaching it from the instance and deleting it from the info cache [ 788.255538] env[70013]: DEBUG nova.network.neutron [req-46ca0ee7-5d32-46a0-9ecf-e00c21530367 req-28024c91-cae9-4426-9a83-ce03f7ad2609 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.426691] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230774, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073252} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.426691] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 788.426691] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe3e8da-8ef9-401f-9632-c09c6c369067 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.459411] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] cfe1478c-53f2-4df8-99a8-ab23209f5306/cfe1478c-53f2-4df8-99a8-ab23209f5306.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.459802] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13822e0a-45cd-4a9b-8849-11928a685869 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.489074] env[70013]: INFO nova.compute.manager [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Took 43.94 seconds to build instance. [ 788.495043] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 788.495043] env[70013]: value = "task-4230775" [ 788.495043] env[70013]: _type = "Task" [ 788.495043] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.504944] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230775, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.611668] env[70013]: INFO nova.compute.manager [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] [instance: 8317b86e-1408-484c-822a-3b2f248c3a4d] Took 1.03 seconds to deallocate network for instance. [ 788.725811] env[70013]: DEBUG nova.network.neutron [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.757992] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1672e1af-dbd3-45ee-92df-e8bb399efade {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.768523] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc4a049-0eba-4353-9668-28d50f7f6c05 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.785917] env[70013]: ERROR nova.scheduler.client.report [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [req-3f4b98ef-c6d3-4b04-a46d-559db195cfde] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3f4b98ef-c6d3-4b04-a46d-559db195cfde"}]} [ 788.786308] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.786898] env[70013]: ERROR nova.compute.manager [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 788.786898] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Traceback (most recent call last): [ 788.786898] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 788.786898] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] yield [ 788.786898] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 788.786898] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] self.set_inventory_for_provider( [ 788.786898] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 788.786898] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3f4b98ef-c6d3-4b04-a46d-559db195cfde"}]} [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] During handling of the above exception, another exception occurred: [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Traceback (most recent call last): [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] with self.rt.instance_claim(context, instance, node, allocs, [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 788.787137] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] return f(*args, **kwargs) [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] self._update(elevated, cn) [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] self._update_to_placement(context, compute_node, startup) [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] return attempt.get(self._wrap_exception) [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] six.reraise(self.value[0], self.value[1], self.value[2]) [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] raise value [ 788.787394] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] self.reportclient.update_from_provider_tree( [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] with catch_all(pd.uuid): [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] self.gen.throw(typ, value, traceback) [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] raise exception.ResourceProviderSyncFailed() [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 788.787781] env[70013]: ERROR nova.compute.manager [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] [ 788.788034] env[70013]: DEBUG nova.compute.utils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 788.789548] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.591s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.789764] env[70013]: DEBUG nova.objects.instance [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lazy-loading 'resources' on Instance uuid 63c3b57c-022a-4eee-b215-6dd89142e659 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.792143] env[70013]: DEBUG nova.compute.manager [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Build of instance b0aad98e-d56c-403c-b075-ddd1634516fb was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 788.792569] env[70013]: DEBUG nova.compute.manager [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 788.792799] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Acquiring lock "refresh_cache-b0aad98e-d56c-403c-b075-ddd1634516fb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.792946] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Acquired lock "refresh_cache-b0aad98e-d56c-403c-b075-ddd1634516fb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.793118] env[70013]: DEBUG nova.network.neutron [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.814446] env[70013]: DEBUG nova.compute.manager [req-46ca0ee7-5d32-46a0-9ecf-e00c21530367 req-28024c91-cae9-4426-9a83-ce03f7ad2609 service nova] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Detach interface failed, port_id=49f610ba-b8a0-4bd6-b0b9-3f15b2635bcd, reason: Instance 85488f5e-bd8a-4bcf-8420-744c54176c2e could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 788.991487] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f8286511-e563-4209-8657-5544a6a46b65 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "67b90fec-a6fd-471d-a425-82bc63de572a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.075s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.009019] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230775, 'name': ReconfigVM_Task, 'duration_secs': 0.31703} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.009486] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Reconfigured VM instance instance-0000001a to attach disk [datastore1] cfe1478c-53f2-4df8-99a8-ab23209f5306/cfe1478c-53f2-4df8-99a8-ab23209f5306.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.010409] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3efe3a20-fad0-4aa4-84f0-c5560e479f78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.022022] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 789.022022] env[70013]: value = "task-4230776" [ 789.022022] env[70013]: _type = "Task" [ 789.022022] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.032640] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230776, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.229422] env[70013]: INFO nova.compute.manager [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Took 1.32 seconds to deallocate network for instance. [ 789.319176] env[70013]: DEBUG nova.network.neutron [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.328339] env[70013]: DEBUG nova.scheduler.client.report [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 789.351396] env[70013]: DEBUG nova.scheduler.client.report [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 789.351627] env[70013]: DEBUG nova.compute.provider_tree [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 789.366239] env[70013]: DEBUG nova.scheduler.client.report [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 789.406982] env[70013]: DEBUG nova.scheduler.client.report [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 789.428865] env[70013]: DEBUG nova.network.neutron [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.500538] env[70013]: DEBUG nova.compute.manager [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 789.542160] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230776, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.608949] env[70013]: DEBUG nova.compute.manager [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Received event network-changed-9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 789.610778] env[70013]: DEBUG nova.compute.manager [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Refreshing instance network info cache due to event network-changed-9e2b8cbf-ea68-4961-ae17-64f6753e3974. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 789.611130] env[70013]: DEBUG oslo_concurrency.lockutils [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] Acquiring lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.611294] env[70013]: DEBUG oslo_concurrency.lockutils [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] Acquired lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.612260] env[70013]: DEBUG nova.network.neutron [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Refreshing network info cache for port 9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.648358] env[70013]: INFO nova.scheduler.client.report [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Deleted allocations for instance 8317b86e-1408-484c-822a-3b2f248c3a4d [ 789.741066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 789.932225] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Releasing lock "refresh_cache-b0aad98e-d56c-403c-b075-ddd1634516fb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.932500] env[70013]: DEBUG nova.compute.manager [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 789.932693] env[70013]: DEBUG nova.compute.manager [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 789.932858] env[70013]: DEBUG nova.network.neutron [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.974441] env[70013]: DEBUG nova.network.neutron [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.016774] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f010a4d-f540-4b84-9e67-288147507453 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.032726] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f0db65-f7eb-4786-94ef-edba7bc8c0c6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.037488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 790.042503] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230776, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.073587] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9a241b-aba2-4df8-83d2-07aa496ad0ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.086360] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e202b10-5215-43e6-af10-d5f3cb247da9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.103128] env[70013]: DEBUG nova.compute.provider_tree [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 790.164049] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd2ffdbc-c90b-4750-841b-0bb80f7c77ae tempest-ServerTagsTestJSON-143473865 tempest-ServerTagsTestJSON-143473865-project-member] Lock "8317b86e-1408-484c-822a-3b2f248c3a4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.744s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 790.479128] env[70013]: DEBUG nova.network.neutron [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.529178] env[70013]: DEBUG nova.network.neutron [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updated VIF entry in instance network info cache for port 9e2b8cbf-ea68-4961-ae17-64f6753e3974. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 790.529178] env[70013]: DEBUG nova.network.neutron [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updating instance_info_cache with network_info: [{"id": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "address": "fa:16:3e:93:5e:8d", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e2b8cbf-ea", "ovs_interfaceid": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.540992] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230776, 'name': Rename_Task, 'duration_secs': 1.160455} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.544351] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 790.544351] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f60a384-be4b-499b-8b84-268a61b223d8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.550594] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 790.550594] env[70013]: value = "task-4230777" [ 790.550594] env[70013]: _type = "Task" [ 790.550594] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.566101] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.637023] env[70013]: ERROR nova.scheduler.client.report [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [req-c8a0f138-d790-4bb5-b577-753219c6e400] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c8a0f138-d790-4bb5-b577-753219c6e400"}]} [ 790.637023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 790.637379] env[70013]: ERROR nova.compute.manager [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 790.637379] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Traceback (most recent call last): [ 790.637379] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 790.637379] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] yield [ 790.637379] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 790.637379] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self.set_inventory_for_provider( [ 790.637379] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 790.637379] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c8a0f138-d790-4bb5-b577-753219c6e400"}]} [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] During handling of the above exception, another exception occurred: [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Traceback (most recent call last): [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self._delete_instance(context, instance, bdms) [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 790.637652] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self._complete_deletion(context, instance) [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self._update_resource_tracker(context, instance) [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self.rt.update_usage(context, instance, instance.node) [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] return f(*args, **kwargs) [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self._update(context.elevated(), self.compute_nodes[nodename]) [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self._update_to_placement(context, compute_node, startup) [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 790.637923] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] return attempt.get(self._wrap_exception) [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] six.reraise(self.value[0], self.value[1], self.value[2]) [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] raise value [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self.reportclient.update_from_provider_tree( [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] with catch_all(pd.uuid): [ 790.638299] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 790.638630] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] self.gen.throw(typ, value, traceback) [ 790.638630] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 790.638630] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] raise exception.ResourceProviderSyncFailed() [ 790.638630] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 790.638630] env[70013]: ERROR nova.compute.manager [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] [ 790.643166] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 28.068s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 790.667319] env[70013]: DEBUG nova.compute.manager [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 790.982600] env[70013]: INFO nova.compute.manager [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] [instance: b0aad98e-d56c-403c-b075-ddd1634516fb] Took 1.05 seconds to deallocate network for instance. [ 791.035866] env[70013]: DEBUG oslo_concurrency.lockutils [req-902f0dfc-cd29-4674-866a-6d0efb9a160d req-427c1a7f-4382-478f-b903-3b1eb376a7f6 service nova] Releasing lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.066867] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230777, 'name': PowerOnVM_Task} progress is 90%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.145982] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.393s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.172187] env[70013]: DEBUG nova.scheduler.client.report [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 791.201790] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.202810] env[70013]: DEBUG nova.scheduler.client.report [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 791.203027] env[70013]: DEBUG nova.compute.provider_tree [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 791.217237] env[70013]: DEBUG nova.scheduler.client.report [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 791.250455] env[70013]: DEBUG nova.scheduler.client.report [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 791.574526] env[70013]: DEBUG oslo_vmware.api [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230777, 'name': PowerOnVM_Task, 'duration_secs': 0.552979} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.574813] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 791.575243] env[70013]: INFO nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Took 9.56 seconds to spawn the instance on the hypervisor. [ 791.575243] env[70013]: DEBUG nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 791.576265] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bac439-c919-4e00-98f2-c97b8fd00438 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.863149] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2aa95e3-3783-4899-bd1f-722d35304331 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.871575] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498b8c17-7f7f-4e28-8681-0cf18bcd8836 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.913483] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57cc5da8-4e4c-4d1b-890c-5bc8f48801bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.918219] env[70013]: DEBUG nova.compute.manager [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Received event network-changed-9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 791.918584] env[70013]: DEBUG nova.compute.manager [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Refreshing instance network info cache due to event network-changed-9e2b8cbf-ea68-4961-ae17-64f6753e3974. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 791.918920] env[70013]: DEBUG oslo_concurrency.lockutils [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] Acquiring lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.919232] env[70013]: DEBUG oslo_concurrency.lockutils [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] Acquired lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 791.919537] env[70013]: DEBUG nova.network.neutron [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Refreshing network info cache for port 9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.928570] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed8f9d3-0631-4b37-a2b8-18bad379ce64 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.949218] env[70013]: DEBUG nova.compute.provider_tree [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.032258] env[70013]: INFO nova.scheduler.client.report [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Deleted allocations for instance b0aad98e-d56c-403c-b075-ddd1634516fb [ 792.109831] env[70013]: INFO nova.compute.manager [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Took 44.91 seconds to build instance. [ 792.478526] env[70013]: ERROR nova.scheduler.client.report [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [req-121208ca-5a86-460a-bc44-3f8192936358] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-121208ca-5a86-460a-bc44-3f8192936358"}]} [ 792.478921] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 1.838s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.479649] env[70013]: ERROR nova.compute.manager [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Confirm resize failed on source host cpu-1. Resource allocations in the placement service will be removed regardless because the instance is now on the destination host cpu-1. You can try hard rebooting the instance to correct its state.: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 792.479649] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Traceback (most recent call last): [ 792.479649] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 792.479649] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] yield [ 792.479649] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 792.479649] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.set_inventory_for_provider( [ 792.479649] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 792.479649] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-121208ca-5a86-460a-bc44-3f8192936358"}]} [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] During handling of the above exception, another exception occurred: [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Traceback (most recent call last): [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/manager.py", line 5279, in do_confirm_resize [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._confirm_resize( [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/manager.py", line 5364, in _confirm_resize [ 792.479905] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.rt.drop_move_claim_at_source(context, instance, migration) [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] return f(*args, **kwargs) [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 636, in drop_move_claim_at_source [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._drop_move_claim( [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 707, in _drop_move_claim [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._update(ctxt, self.compute_nodes[nodename]) [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._update_to_placement(context, compute_node, startup) [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 792.480154] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] return attempt.get(self._wrap_exception) [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] six.reraise(self.value[0], self.value[1], self.value[2]) [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] raise value [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.reportclient.update_from_provider_tree( [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] with catch_all(pd.uuid): [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.gen.throw(typ, value, traceback) [ 792.480457] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 792.480939] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] raise exception.ResourceProviderSyncFailed() [ 792.480939] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 792.480939] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] [ 792.482403] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.708s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.483984] env[70013]: INFO nova.compute.claims [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.541937] env[70013]: DEBUG oslo_concurrency.lockutils [None req-205ca9d8-43aa-4927-95d5-ab0bc61f2bd7 tempest-FloatingIPsAssociationNegativeTestJSON-110477298 tempest-FloatingIPsAssociationNegativeTestJSON-110477298-project-member] Lock "b0aad98e-d56c-403c-b075-ddd1634516fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.726s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.572782] env[70013]: INFO nova.scheduler.client.report [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Deleted allocation for migration 2b86447b-3ba9-457a-94d4-7269f0b254ab [ 792.616597] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b726d792-fd0c-4941-a734-fa707e8a7f36 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.355s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.658198] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.693495] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f7d5b-9053-4aea-674f-bce41eb4659a/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 792.694902] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de96990b-6ae9-4254-a2a3-2b9882e26158 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.704543] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f7d5b-9053-4aea-674f-bce41eb4659a/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 792.704543] env[70013]: ERROR oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f7d5b-9053-4aea-674f-bce41eb4659a/disk-0.vmdk due to incomplete transfer. [ 792.704543] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bfad4da5-4792-42c7-90ce-f19a9af9308e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.714312] env[70013]: DEBUG oslo_vmware.rw_handles [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f7d5b-9053-4aea-674f-bce41eb4659a/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 792.714594] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Uploaded image d4b5d8a9-aa58-4102-b923-43a5fc4d950f to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 792.716872] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 792.718121] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-509e5ebf-18c2-44a1-857c-75b53aa8a3ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.727853] env[70013]: DEBUG nova.network.neutron [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updated VIF entry in instance network info cache for port 9e2b8cbf-ea68-4961-ae17-64f6753e3974. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 792.728038] env[70013]: DEBUG nova.network.neutron [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updating instance_info_cache with network_info: [{"id": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "address": "fa:16:3e:93:5e:8d", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e2b8cbf-ea", "ovs_interfaceid": "9e2b8cbf-ea68-4961-ae17-64f6753e3974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.731384] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 792.731384] env[70013]: value = "task-4230778" [ 792.731384] env[70013]: _type = "Task" [ 792.731384] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.741724] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230778, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.047241] env[70013]: DEBUG nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 793.085068] env[70013]: ERROR nova.compute.manager [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 793.085068] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Traceback (most recent call last): [ 793.085068] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 793.085068] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] yield [ 793.085068] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 793.085068] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.set_inventory_for_provider( [ 793.085068] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 793.085068] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-121208ca-5a86-460a-bc44-3f8192936358"}]} [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] During handling of the above exception, another exception occurred: [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Traceback (most recent call last): [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/manager.py", line 11501, in _error_out_instance_on_exception [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] yield [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/manager.py", line 5285, in do_confirm_resize [ 793.085381] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] with excutils.save_and_reraise_exception(logger=LOG): [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.force_reraise() [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] raise self.value [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/manager.py", line 5279, in do_confirm_resize [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._confirm_resize( [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/manager.py", line 5364, in _confirm_resize [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.rt.drop_move_claim_at_source(context, instance, migration) [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] return f(*args, **kwargs) [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 636, in drop_move_claim_at_source [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._drop_move_claim( [ 793.085609] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 707, in _drop_move_claim [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._update(ctxt, self.compute_nodes[nodename]) [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self._update_to_placement(context, compute_node, startup) [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] return attempt.get(self._wrap_exception) [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] six.reraise(self.value[0], self.value[1], self.value[2]) [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] raise value [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 793.085885] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.reportclient.update_from_provider_tree( [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] with catch_all(pd.uuid): [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] self.gen.throw(typ, value, traceback) [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] raise exception.ResourceProviderSyncFailed() [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 793.086240] env[70013]: ERROR nova.compute.manager [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] [ 793.123602] env[70013]: DEBUG nova.compute.manager [None req-e668e455-0474-4b05-80d7-881ae19e3a8c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: d3d5b3ff-9ac4-4c6b-b9d9-daa306275c5a] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 793.232741] env[70013]: DEBUG oslo_concurrency.lockutils [req-03ae797e-b164-4855-a2f3-7bbbf30d8c60 req-611bbb67-2d40-4f56-b40f-bba965c55610 service nova] Releasing lock "refresh_cache-67b90fec-a6fd-471d-a425-82bc63de572a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.245597] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230778, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.518779] env[70013]: DEBUG nova.scheduler.client.report [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 793.543780] env[70013]: DEBUG nova.scheduler.client.report [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 793.543879] env[70013]: DEBUG nova.compute.provider_tree [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 793.570628] env[70013]: DEBUG nova.scheduler.client.report [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 793.594115] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 34.838s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.601182] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.608435] env[70013]: DEBUG nova.scheduler.client.report [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 793.629304] env[70013]: DEBUG nova.compute.manager [None req-e668e455-0474-4b05-80d7-881ae19e3a8c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: d3d5b3ff-9ac4-4c6b-b9d9-daa306275c5a] Instance disappeared before build. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2446}} [ 793.700835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "477bf0d8-e988-43e9-a573-994a0d08df0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.701208] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "477bf0d8-e988-43e9-a573-994a0d08df0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.750476] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230778, 'name': Destroy_Task, 'duration_secs': 0.976374} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.751177] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Destroyed the VM [ 793.751177] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 793.752041] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-55f1b045-719e-425d-a64e-591837660f39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.759844] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 793.759844] env[70013]: value = "task-4230779" [ 793.759844] env[70013]: _type = "Task" [ 793.759844] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.773819] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230779, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.147805] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e668e455-0474-4b05-80d7-881ae19e3a8c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "d3d5b3ff-9ac4-4c6b-b9d9-daa306275c5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.020s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.221432] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a2c4cb-1d86-4b87-a084-a891d421a308 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.235027] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bce1751-0bf2-4849-bbfd-72d057c4beee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.277903] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a16b6f4-2f37-4a02-8795-973bc4a9e303 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.289391] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230779, 'name': RemoveSnapshot_Task} progress is 26%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.294934] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26911fc2-838f-4e8a-bb3e-2d0fefb04902 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.312289] env[70013]: DEBUG nova.compute.provider_tree [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 794.618313] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 794.618806] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 794.650750] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 794.784147] env[70013]: DEBUG oslo_vmware.api [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230779, 'name': RemoveSnapshot_Task, 'duration_secs': 0.809973} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.784653] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 794.785027] env[70013]: INFO nova.compute.manager [None req-0036a97b-c392-44cb-ab3d-0952ef489562 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Took 14.90 seconds to snapshot the instance on the hypervisor. [ 794.849620] env[70013]: ERROR nova.scheduler.client.report [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [req-6ca10dea-6ebc-40b3-b338-79f6dd0fe3a0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6ca10dea-6ebc-40b3-b338-79f6dd0fe3a0"}]} [ 794.849620] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.849723] env[70013]: ERROR nova.compute.manager [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 794.849723] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Traceback (most recent call last): [ 794.849723] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 794.849723] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] yield [ 794.849723] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 794.849723] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] self.set_inventory_for_provider( [ 794.849723] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 794.849723] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6ca10dea-6ebc-40b3-b338-79f6dd0fe3a0"}]} [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] During handling of the above exception, another exception occurred: [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Traceback (most recent call last): [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] with self.rt.instance_claim(context, instance, node, allocs, [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 794.849929] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] return f(*args, **kwargs) [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] self._update(elevated, cn) [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] self._update_to_placement(context, compute_node, startup) [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] return attempt.get(self._wrap_exception) [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] six.reraise(self.value[0], self.value[1], self.value[2]) [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] raise value [ 794.850254] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] self.reportclient.update_from_provider_tree( [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] with catch_all(pd.uuid): [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] self.gen.throw(typ, value, traceback) [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] raise exception.ResourceProviderSyncFailed() [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 794.850594] env[70013]: ERROR nova.compute.manager [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] [ 794.850906] env[70013]: DEBUG nova.compute.utils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 794.854147] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.004s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 794.854147] env[70013]: INFO nova.compute.claims [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.857657] env[70013]: DEBUG nova.compute.manager [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Build of instance 566c6b47-34ce-4d53-baad-f31bf08ad3a5 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 794.859405] env[70013]: DEBUG nova.compute.manager [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 794.860377] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Acquiring lock "refresh_cache-566c6b47-34ce-4d53-baad-f31bf08ad3a5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.861183] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Acquired lock "refresh_cache-566c6b47-34ce-4d53-baad-f31bf08ad3a5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.862918] env[70013]: DEBUG nova.network.neutron [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server [None req-f40a99e1-34ba-4504-a123-e994212b1443 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server yield [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-121208ca-5a86-460a-bc44-3f8192936358"}]} [ 795.116083] env[70013]: ERROR oslo_messaging.rpc.server [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 795.116441] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 146, in decorated_function [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5308, in confirm_resize [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server do_confirm_resize(context, instance, migration) [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5285, in do_confirm_resize [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(logger=LOG): [ 795.116853] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5279, in do_confirm_resize [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server self._confirm_resize( [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5364, in _confirm_resize [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server self.rt.drop_move_claim_at_source(context, instance, migration) [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 636, in drop_move_claim_at_source [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server self._drop_move_claim( [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 707, in _drop_move_claim [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server self._update(ctxt, self.compute_nodes[nodename]) [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 795.117343] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 795.117750] env[70013]: ERROR oslo_messaging.rpc.server [ 795.179654] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.394541] env[70013]: DEBUG nova.network.neutron [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.496405] env[70013]: DEBUG nova.network.neutron [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.755636] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "cfe1478c-53f2-4df8-99a8-ab23209f5306" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.755993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.756167] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "cfe1478c-53f2-4df8-99a8-ab23209f5306-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.756362] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.756548] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.759368] env[70013]: INFO nova.compute.manager [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Terminating instance [ 795.892475] env[70013]: DEBUG nova.scheduler.client.report [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 795.919849] env[70013]: DEBUG nova.scheduler.client.report [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 795.919849] env[70013]: DEBUG nova.compute.provider_tree [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 795.932769] env[70013]: DEBUG nova.scheduler.client.report [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 795.956690] env[70013]: DEBUG nova.scheduler.client.report [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 796.000193] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Releasing lock "refresh_cache-566c6b47-34ce-4d53-baad-f31bf08ad3a5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 796.000497] env[70013]: DEBUG nova.compute.manager [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 796.000695] env[70013]: DEBUG nova.compute.manager [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 796.000868] env[70013]: DEBUG nova.network.neutron [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.026718] env[70013]: DEBUG nova.network.neutron [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.266867] env[70013]: DEBUG nova.compute.manager [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 796.267136] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.268102] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99283b36-36a2-46f2-a29e-97cdb20cddba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.278380] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.278619] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe8fcb86-84f2-4748-a73a-1914e3c66489 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.286932] env[70013]: DEBUG oslo_vmware.api [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 796.286932] env[70013]: value = "task-4230780" [ 796.286932] env[70013]: _type = "Task" [ 796.286932] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.301938] env[70013]: DEBUG oslo_vmware.api [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230780, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.510066] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea97f4ed-b145-4a17-8db8-a13f3b45ec13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.518300] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5385edf2-4359-40c7-abe5-141ce15ec5d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.554030] env[70013]: DEBUG nova.network.neutron [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.557306] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0a1d9d-59e7-45b5-8456-4bb0251f65cf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.568919] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6093e6-3e2a-41a0-a16d-47927a97b894 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.584807] env[70013]: DEBUG nova.compute.provider_tree [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.803378] env[70013]: DEBUG oslo_vmware.api [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230780, 'name': PowerOffVM_Task, 'duration_secs': 0.20847} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.804497] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.804497] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 796.804497] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b024e67-96c5-4d89-b2e7-e8f2eeea214c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.841793] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.842245] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.878150] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 796.878396] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 796.879088] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Deleting the datastore file [datastore1] cfe1478c-53f2-4df8-99a8-ab23209f5306 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.879715] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7666ffc-12de-4c4e-bfbe-74a28baccca1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.889982] env[70013]: DEBUG oslo_vmware.api [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for the task: (returnval){ [ 796.889982] env[70013]: value = "task-4230782" [ 796.889982] env[70013]: _type = "Task" [ 796.889982] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.901666] env[70013]: DEBUG oslo_vmware.api [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230782, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.058325] env[70013]: INFO nova.compute.manager [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] [instance: 566c6b47-34ce-4d53-baad-f31bf08ad3a5] Took 1.06 seconds to deallocate network for instance. [ 797.088247] env[70013]: DEBUG nova.scheduler.client.report [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 797.381147] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "f12b8575-5082-4be9-9bf5-f4279860d19d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.381501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.381744] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "f12b8575-5082-4be9-9bf5-f4279860d19d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.381947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.382141] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.385472] env[70013]: INFO nova.compute.manager [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Terminating instance [ 797.407176] env[70013]: DEBUG oslo_vmware.api [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Task: {'id': task-4230782, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140603} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.407346] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.407762] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.407762] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.408176] env[70013]: INFO nova.compute.manager [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Took 1.14 seconds to destroy the instance on the hypervisor. [ 797.408176] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 797.408706] env[70013]: DEBUG nova.compute.manager [-] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 797.408706] env[70013]: DEBUG nova.network.neutron [-] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.600267] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.745s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.600267] env[70013]: DEBUG nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 797.602997] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.674s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.604547] env[70013]: INFO nova.compute.claims [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.797733] env[70013]: DEBUG nova.compute.manager [req-d3d79217-49b1-4e46-b0a8-2149be38f2fb req-bf1c7c23-3fba-4bd0-8e63-f86da66d2ee0 service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Received event network-vif-deleted-42a0109b-8ff8-4977-b5ce-f0caa32df87e {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 797.797948] env[70013]: INFO nova.compute.manager [req-d3d79217-49b1-4e46-b0a8-2149be38f2fb req-bf1c7c23-3fba-4bd0-8e63-f86da66d2ee0 service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Neutron deleted interface 42a0109b-8ff8-4977-b5ce-f0caa32df87e; detaching it from the instance and deleting it from the info cache [ 797.798126] env[70013]: DEBUG nova.network.neutron [req-d3d79217-49b1-4e46-b0a8-2149be38f2fb req-bf1c7c23-3fba-4bd0-8e63-f86da66d2ee0 service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.890347] env[70013]: DEBUG nova.compute.manager [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 797.890627] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.891538] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38027c54-8596-4d3d-b736-991cd18eb29f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.899979] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.900257] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a20c8f0-3538-48a2-99db-ff23df2ed901 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.907886] env[70013]: DEBUG oslo_vmware.api [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 797.907886] env[70013]: value = "task-4230783" [ 797.907886] env[70013]: _type = "Task" [ 797.907886] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.918216] env[70013]: DEBUG oslo_vmware.api [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.088749] env[70013]: INFO nova.scheduler.client.report [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Deleted allocations for instance 566c6b47-34ce-4d53-baad-f31bf08ad3a5 [ 798.105885] env[70013]: DEBUG nova.compute.utils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 798.107294] env[70013]: DEBUG nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 798.108477] env[70013]: DEBUG nova.network.neutron [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 798.193098] env[70013]: DEBUG nova.policy [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19becdb6b16d4f9eac99aaf6ece00c7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89e4599a0f9342b79d5eb9b2eebb51a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 798.195239] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.195955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.281563] env[70013]: DEBUG nova.network.neutron [-] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.303394] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-360dab40-1015-47c4-8bc6-f7242028212f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.315071] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc189201-cd6d-46d3-9119-d34b3fa86c84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.359503] env[70013]: DEBUG nova.compute.manager [req-d3d79217-49b1-4e46-b0a8-2149be38f2fb req-bf1c7c23-3fba-4bd0-8e63-f86da66d2ee0 service nova] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Detach interface failed, port_id=42a0109b-8ff8-4977-b5ce-f0caa32df87e, reason: Instance cfe1478c-53f2-4df8-99a8-ab23209f5306 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 798.420102] env[70013]: DEBUG oslo_vmware.api [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230783, 'name': PowerOffVM_Task, 'duration_secs': 0.231964} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.420690] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 798.420957] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 798.421333] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9eafed51-1eeb-4651-8b10-52ff98f160a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.497489] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 798.497847] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 798.498108] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Deleting the datastore file [datastore2] f12b8575-5082-4be9-9bf5-f4279860d19d {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.498431] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af334f3e-eef8-44c1-a80f-956aa0d19dc2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.507406] env[70013]: DEBUG oslo_vmware.api [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for the task: (returnval){ [ 798.507406] env[70013]: value = "task-4230785" [ 798.507406] env[70013]: _type = "Task" [ 798.507406] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.519528] env[70013]: DEBUG oslo_vmware.api [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.598635] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ade4c7b3-fd9c-4750-80bf-49995968f81f tempest-ImagesOneServerNegativeTestJSON-2088258160 tempest-ImagesOneServerNegativeTestJSON-2088258160-project-member] Lock "566c6b47-34ce-4d53-baad-f31bf08ad3a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.036s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.612809] env[70013]: DEBUG nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 798.624682] env[70013]: DEBUG nova.network.neutron [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Successfully created port: c7fc6ffa-a891-4805-adfe-0a1ed770b245 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.784388] env[70013]: INFO nova.compute.manager [-] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Took 1.38 seconds to deallocate network for instance. [ 799.021941] env[70013]: DEBUG oslo_vmware.api [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Task: {'id': task-4230785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142995} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.024825] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.025291] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 799.025457] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.025635] env[70013]: INFO nova.compute.manager [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 799.025953] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 799.027530] env[70013]: DEBUG nova.compute.manager [-] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 799.027530] env[70013]: DEBUG nova.network.neutron [-] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 799.105825] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 799.195018] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7703db3b-f0be-47d9-9809-8687fff0e645 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.204503] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68d71b1-9fc2-4153-9d92-de9d419b92df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.243727] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9cfda8-2d33-4cdf-a74e-021f52a9bc9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.254882] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29696a2a-5a1b-4d28-90b4-75a9f8e6a12b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.269972] env[70013]: DEBUG nova.compute.provider_tree [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.292724] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.619959] env[70013]: DEBUG nova.compute.manager [req-78c942d8-3136-49fe-91ab-c45e3cdd3dc0 req-8fdd410c-d0f7-4283-9ceb-4d3978c5cbb9 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Received event network-vif-deleted-3e878689-ac09-469d-8490-a73b6e02d403 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 799.621905] env[70013]: INFO nova.compute.manager [req-78c942d8-3136-49fe-91ab-c45e3cdd3dc0 req-8fdd410c-d0f7-4283-9ceb-4d3978c5cbb9 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Neutron deleted interface 3e878689-ac09-469d-8490-a73b6e02d403; detaching it from the instance and deleting it from the info cache [ 799.622569] env[70013]: DEBUG nova.network.neutron [req-78c942d8-3136-49fe-91ab-c45e3cdd3dc0 req-8fdd410c-d0f7-4283-9ceb-4d3978c5cbb9 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.632674] env[70013]: DEBUG nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 799.651452] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.667717] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 799.667925] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.668165] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 799.668439] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.668653] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 799.668877] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 799.669255] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 799.669519] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 799.669786] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 799.670051] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 799.670338] env[70013]: DEBUG nova.virt.hardware [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 799.672058] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab12806-94f4-4de1-b363-0b1958a1aea7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.686854] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214a1af7-c61b-4868-ac2f-ef8ef3f1b6e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.774561] env[70013]: DEBUG nova.scheduler.client.report [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 800.018509] env[70013]: DEBUG nova.network.neutron [-] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.133042] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4cc02b2-de3b-48c4-8870-f292c8522740 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.144519] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131f6fd4-1b00-4bc5-909e-75bfc95fd9fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.181453] env[70013]: DEBUG nova.compute.manager [req-78c942d8-3136-49fe-91ab-c45e3cdd3dc0 req-8fdd410c-d0f7-4283-9ceb-4d3978c5cbb9 service nova] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Detach interface failed, port_id=3e878689-ac09-469d-8490-a73b6e02d403, reason: Instance f12b8575-5082-4be9-9bf5-f4279860d19d could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 800.195395] env[70013]: DEBUG nova.compute.manager [req-e06f44f6-2a78-44e7-bc51-6ba8c8d11e89 req-a88a8a24-3e32-4a9c-9cbd-bd37ee0d3b49 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Received event network-vif-plugged-c7fc6ffa-a891-4805-adfe-0a1ed770b245 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 800.195618] env[70013]: DEBUG oslo_concurrency.lockutils [req-e06f44f6-2a78-44e7-bc51-6ba8c8d11e89 req-a88a8a24-3e32-4a9c-9cbd-bd37ee0d3b49 service nova] Acquiring lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 800.195832] env[70013]: DEBUG oslo_concurrency.lockutils [req-e06f44f6-2a78-44e7-bc51-6ba8c8d11e89 req-a88a8a24-3e32-4a9c-9cbd-bd37ee0d3b49 service nova] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.196007] env[70013]: DEBUG oslo_concurrency.lockutils [req-e06f44f6-2a78-44e7-bc51-6ba8c8d11e89 req-a88a8a24-3e32-4a9c-9cbd-bd37ee0d3b49 service nova] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.196517] env[70013]: DEBUG nova.compute.manager [req-e06f44f6-2a78-44e7-bc51-6ba8c8d11e89 req-a88a8a24-3e32-4a9c-9cbd-bd37ee0d3b49 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] No waiting events found dispatching network-vif-plugged-c7fc6ffa-a891-4805-adfe-0a1ed770b245 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 800.196779] env[70013]: WARNING nova.compute.manager [req-e06f44f6-2a78-44e7-bc51-6ba8c8d11e89 req-a88a8a24-3e32-4a9c-9cbd-bd37ee0d3b49 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Received unexpected event network-vif-plugged-c7fc6ffa-a891-4805-adfe-0a1ed770b245 for instance with vm_state building and task_state spawning. [ 800.284807] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.682s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.288027] env[70013]: DEBUG nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 800.289546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.374s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.291619] env[70013]: INFO nova.compute.claims [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.380800] env[70013]: DEBUG nova.network.neutron [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Successfully updated port: c7fc6ffa-a891-4805-adfe-0a1ed770b245 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.520739] env[70013]: INFO nova.compute.manager [-] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Took 1.49 seconds to deallocate network for instance. [ 800.797735] env[70013]: DEBUG nova.compute.utils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 800.803235] env[70013]: DEBUG nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 800.803235] env[70013]: DEBUG nova.network.neutron [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.876946] env[70013]: DEBUG nova.policy [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd357d73286774d2cbe0ab68494c7c57b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75ec7d6e9ea54c06bd3b7605f16b391a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 800.889199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "refresh_cache-2ee62ad3-0125-47dc-b163-7d15b6a17c8a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.889199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquired lock "refresh_cache-2ee62ad3-0125-47dc-b163-7d15b6a17c8a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 800.889434] env[70013]: DEBUG nova.network.neutron [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.028405] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.304323] env[70013]: DEBUG nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 801.395460] env[70013]: DEBUG nova.network.neutron [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Successfully created port: 7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.432856] env[70013]: DEBUG nova.network.neutron [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.715393] env[70013]: DEBUG nova.network.neutron [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Updating instance_info_cache with network_info: [{"id": "c7fc6ffa-a891-4805-adfe-0a1ed770b245", "address": "fa:16:3e:03:d5:8c", "network": {"id": "5b2ea88e-52ce-4872-933d-4e988636a73d", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1918711081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4599a0f9342b79d5eb9b2eebb51a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7fc6ffa-a8", "ovs_interfaceid": "c7fc6ffa-a891-4805-adfe-0a1ed770b245", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.978959] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f24c48-00df-44b2-b549-077d4970d2db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.989400] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46b0338-35dd-4c85-bec1-3b13c4a89847 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.022088] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4d0143-5048-4015-83f5-c57f000df04a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.032279] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304ff078-dc87-475b-a814-6f6fb2e12bcc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.049046] env[70013]: DEBUG nova.compute.provider_tree [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.218188] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Releasing lock "refresh_cache-2ee62ad3-0125-47dc-b163-7d15b6a17c8a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 802.219140] env[70013]: DEBUG nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Instance network_info: |[{"id": "c7fc6ffa-a891-4805-adfe-0a1ed770b245", "address": "fa:16:3e:03:d5:8c", "network": {"id": "5b2ea88e-52ce-4872-933d-4e988636a73d", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1918711081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4599a0f9342b79d5eb9b2eebb51a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7fc6ffa-a8", "ovs_interfaceid": "c7fc6ffa-a891-4805-adfe-0a1ed770b245", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 802.219288] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:d5:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69744f59-ecac-4b0b-831e-82a274d7acbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7fc6ffa-a891-4805-adfe-0a1ed770b245', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.230841] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Creating folder: Project (89e4599a0f9342b79d5eb9b2eebb51a0). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.231333] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ea20335-4bf2-4879-a17b-63149d3ed942 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.244689] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Created folder: Project (89e4599a0f9342b79d5eb9b2eebb51a0) in parent group-v836999. [ 802.244689] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Creating folder: Instances. Parent ref: group-v837091. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.244930] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04b2cba2-09fd-4a8e-be21-cf071d63a394 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.256059] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Created folder: Instances in parent group-v837091. [ 802.256611] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 802.256872] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.257202] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d668717a-474e-4e55-933e-bd8a8b988ffd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.286167] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.286167] env[70013]: value = "task-4230788" [ 802.286167] env[70013]: _type = "Task" [ 802.286167] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.296195] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230788, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.323046] env[70013]: DEBUG nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 802.349234] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 802.350312] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.350312] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 802.350481] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.350651] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 802.350851] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 802.351123] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 802.351904] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 802.352137] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 802.352363] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 802.352587] env[70013]: DEBUG nova.virt.hardware [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 802.354162] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84768473-a49e-49ed-bba6-d2d96c7edc8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.365499] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e5e6d7-11f8-484a-8f00-8ca556ff2d5c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.551856] env[70013]: DEBUG nova.scheduler.client.report [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 802.559786] env[70013]: DEBUG nova.compute.manager [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Received event network-changed-c7fc6ffa-a891-4805-adfe-0a1ed770b245 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 802.559786] env[70013]: DEBUG nova.compute.manager [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Refreshing instance network info cache due to event network-changed-c7fc6ffa-a891-4805-adfe-0a1ed770b245. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 802.560037] env[70013]: DEBUG oslo_concurrency.lockutils [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] Acquiring lock "refresh_cache-2ee62ad3-0125-47dc-b163-7d15b6a17c8a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.560232] env[70013]: DEBUG oslo_concurrency.lockutils [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] Acquired lock "refresh_cache-2ee62ad3-0125-47dc-b163-7d15b6a17c8a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.560474] env[70013]: DEBUG nova.network.neutron [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Refreshing network info cache for port c7fc6ffa-a891-4805-adfe-0a1ed770b245 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.796983] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230788, 'name': CreateVM_Task, 'duration_secs': 0.338535} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.797220] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.797943] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.798185] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.798547] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 802.798846] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bebc2f3a-04ec-4504-b582-dd09dd4bcd41 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.804630] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 802.804630] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a81fa-e43a-adf2-5824-bfe50f53d65b" [ 802.804630] env[70013]: _type = "Task" [ 802.804630] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.813012] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a81fa-e43a-adf2-5824-bfe50f53d65b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.041243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "789fffd9-6725-4bf6-9144-dd603b0a521f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.041243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.066126] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.777s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.069555] env[70013]: DEBUG nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 803.074394] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.841s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.076219] env[70013]: INFO nova.compute.claims [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.143694] env[70013]: DEBUG nova.network.neutron [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Successfully updated port: 7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.318519] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a81fa-e43a-adf2-5824-bfe50f53d65b, 'name': SearchDatastore_Task, 'duration_secs': 0.012809} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.319088] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.319457] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.320666] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.320666] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 803.320666] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.320853] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38bc3c2d-c10a-40c2-9734-a52b4e5c55c3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.333077] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.333077] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 803.334236] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d13573d-faa8-4a9d-8259-7597184fa363 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.341584] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 803.341584] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527ce38c-216d-8817-bdbc-b97eefb0a494" [ 803.341584] env[70013]: _type = "Task" [ 803.341584] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.353796] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527ce38c-216d-8817-bdbc-b97eefb0a494, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.576377] env[70013]: DEBUG nova.compute.utils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 803.577855] env[70013]: DEBUG nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 803.652634] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-77c12460-9cfa-41c9-a210-238a470d9ccd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.652808] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-77c12460-9cfa-41c9-a210-238a470d9ccd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 803.652985] env[70013]: DEBUG nova.network.neutron [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.672171] env[70013]: DEBUG nova.network.neutron [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Updated VIF entry in instance network info cache for port c7fc6ffa-a891-4805-adfe-0a1ed770b245. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 803.672280] env[70013]: DEBUG nova.network.neutron [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Updating instance_info_cache with network_info: [{"id": "c7fc6ffa-a891-4805-adfe-0a1ed770b245", "address": "fa:16:3e:03:d5:8c", "network": {"id": "5b2ea88e-52ce-4872-933d-4e988636a73d", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1918711081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4599a0f9342b79d5eb9b2eebb51a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7fc6ffa-a8", "ovs_interfaceid": "c7fc6ffa-a891-4805-adfe-0a1ed770b245", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.855439] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527ce38c-216d-8817-bdbc-b97eefb0a494, 'name': SearchDatastore_Task, 'duration_secs': 0.015315} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.856252] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd96eed2-5d17-452c-b394-6262de250d6d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.867689] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 803.867689] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521825a5-1dd2-9b06-7e5e-b73fb5335e1c" [ 803.867689] env[70013]: _type = "Task" [ 803.867689] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.882840] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521825a5-1dd2-9b06-7e5e-b73fb5335e1c, 'name': SearchDatastore_Task, 'duration_secs': 0.010339} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.882840] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.882840] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2ee62ad3-0125-47dc-b163-7d15b6a17c8a/2ee62ad3-0125-47dc-b163-7d15b6a17c8a.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.882840] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb3f7da6-7455-4923-b658-84cf2405b547 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.890980] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 803.890980] env[70013]: value = "task-4230789" [ 803.890980] env[70013]: _type = "Task" [ 803.890980] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.900266] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230789, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.079061] env[70013]: DEBUG nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 804.175524] env[70013]: DEBUG oslo_concurrency.lockutils [req-1025426b-0ec5-491a-b523-fce20104485f req-13f006c4-643a-49c8-bb08-c8cdcaec6eb5 service nova] Releasing lock "refresh_cache-2ee62ad3-0125-47dc-b163-7d15b6a17c8a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 804.238884] env[70013]: DEBUG nova.network.neutron [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.314602] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 804.315104] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.410710] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230789, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.560936] env[70013]: DEBUG nova.network.neutron [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Updating instance_info_cache with network_info: [{"id": "7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3", "address": "fa:16:3e:58:41:6f", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7feff77e-7b", "ovs_interfaceid": "7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.601035] env[70013]: DEBUG nova.compute.manager [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Received event network-vif-plugged-7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 804.601270] env[70013]: DEBUG oslo_concurrency.lockutils [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] Acquiring lock "77c12460-9cfa-41c9-a210-238a470d9ccd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 804.601491] env[70013]: DEBUG oslo_concurrency.lockutils [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.601697] env[70013]: DEBUG oslo_concurrency.lockutils [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.601868] env[70013]: DEBUG nova.compute.manager [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] No waiting events found dispatching network-vif-plugged-7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 804.602047] env[70013]: WARNING nova.compute.manager [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Received unexpected event network-vif-plugged-7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 for instance with vm_state building and task_state spawning. [ 804.602214] env[70013]: DEBUG nova.compute.manager [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Received event network-changed-7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 804.602365] env[70013]: DEBUG nova.compute.manager [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Refreshing instance network info cache due to event network-changed-7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 804.602723] env[70013]: DEBUG oslo_concurrency.lockutils [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] Acquiring lock "refresh_cache-77c12460-9cfa-41c9-a210-238a470d9ccd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.723655] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46fc8d5-74bd-4b47-8d9b-1d7e30fb7829 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.733041] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5057b0d0-c45c-4170-9519-fcc1a4fda7bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.773399] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f652c7-a761-436b-8bf4-062a5bb9b69f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.782555] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c72e258-8635-44e4-9b73-775912891347 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.801126] env[70013]: DEBUG nova.compute.provider_tree [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.905174] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230789, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535324} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.905174] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2ee62ad3-0125-47dc-b163-7d15b6a17c8a/2ee62ad3-0125-47dc-b163-7d15b6a17c8a.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.905439] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.905695] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd09a19c-4c1b-46a2-8f75-3e24d18b84ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.913965] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 804.913965] env[70013]: value = "task-4230790" [ 804.913965] env[70013]: _type = "Task" [ 804.913965] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.925545] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230790, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.063236] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-77c12460-9cfa-41c9-a210-238a470d9ccd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.063638] env[70013]: DEBUG nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Instance network_info: |[{"id": "7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3", "address": "fa:16:3e:58:41:6f", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7feff77e-7b", "ovs_interfaceid": "7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 805.064028] env[70013]: DEBUG oslo_concurrency.lockutils [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] Acquired lock "refresh_cache-77c12460-9cfa-41c9-a210-238a470d9ccd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.064235] env[70013]: DEBUG nova.network.neutron [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Refreshing network info cache for port 7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.065638] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:41:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.077923] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 805.079125] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.079509] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e71bfce4-5b60-4559-b102-f881aff765f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.101378] env[70013]: DEBUG nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 805.109534] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.109534] env[70013]: value = "task-4230791" [ 805.109534] env[70013]: _type = "Task" [ 805.109534] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.122707] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230791, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.141969] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 805.142362] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.142513] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 805.142996] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.143230] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 805.143468] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 805.143690] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 805.143886] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 805.144105] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 805.144315] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 805.144531] env[70013]: DEBUG nova.virt.hardware [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 805.146354] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448059e0-0608-452e-9ce6-8afde369339d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.155469] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084f4f5f-39e7-46df-95fe-7f72046e6c22 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.174348] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.183383] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Creating folder: Project (d90c7119761f4982bb3e5b2a831a83f8). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.183383] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc76c6fa-6cb6-4d3d-bab0-a1c86d0a59e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.196725] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Created folder: Project (d90c7119761f4982bb3e5b2a831a83f8) in parent group-v836999. [ 805.197119] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Creating folder: Instances. Parent ref: group-v837095. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.197423] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18944b21-093c-4e56-8ce9-72dfdc0e5497 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.210180] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Created folder: Instances in parent group-v837095. [ 805.210536] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 805.210779] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.211015] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac33bd1e-fac1-47db-96a7-456abbb83d05 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.234217] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.234217] env[70013]: value = "task-4230794" [ 805.234217] env[70013]: _type = "Task" [ 805.234217] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.244680] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230794, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.308381] env[70013]: DEBUG nova.scheduler.client.report [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 805.425430] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230790, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080041} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.425809] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.427092] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666431c8-2a70-4ac0-aae4-39be754131ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.455498] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 2ee62ad3-0125-47dc-b163-7d15b6a17c8a/2ee62ad3-0125-47dc-b163-7d15b6a17c8a.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.455859] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d0799e3-e94f-4b36-8a26-341dd9132163 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.479530] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 805.479530] env[70013]: value = "task-4230795" [ 805.479530] env[70013]: _type = "Task" [ 805.479530] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.496426] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230795, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.621445] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230791, 'name': CreateVM_Task, 'duration_secs': 0.416612} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.624854] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.625587] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.625763] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.626093] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 805.626646] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e55e2543-78df-4d73-b58f-bb484c6e104e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.631733] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 805.631733] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52129753-31cd-917d-22e9-dd438c54e792" [ 805.631733] env[70013]: _type = "Task" [ 805.631733] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.640669] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52129753-31cd-917d-22e9-dd438c54e792, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.747548] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230794, 'name': CreateVM_Task, 'duration_secs': 0.35537} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.747734] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.748155] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.817872] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.743s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.818303] env[70013]: DEBUG nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 805.821338] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.041s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.821573] env[70013]: DEBUG nova.objects.instance [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lazy-loading 'resources' on Instance uuid 062eff58-95fe-4c9b-a586-7e7434c77adf {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 805.974304] env[70013]: DEBUG nova.network.neutron [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Updated VIF entry in instance network info cache for port 7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 805.974763] env[70013]: DEBUG nova.network.neutron [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Updating instance_info_cache with network_info: [{"id": "7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3", "address": "fa:16:3e:58:41:6f", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7feff77e-7b", "ovs_interfaceid": "7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.995989] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230795, 'name': ReconfigVM_Task, 'duration_secs': 0.416818} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.996315] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 2ee62ad3-0125-47dc-b163-7d15b6a17c8a/2ee62ad3-0125-47dc-b163-7d15b6a17c8a.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.997329] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ac901d7-9a40-4059-9d9a-8f0ddbbacd62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.004782] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 806.004782] env[70013]: value = "task-4230796" [ 806.004782] env[70013]: _type = "Task" [ 806.004782] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.016788] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230796, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.146160] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52129753-31cd-917d-22e9-dd438c54e792, 'name': SearchDatastore_Task, 'duration_secs': 0.009607} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.146509] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.147863] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.147863] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.147863] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.147863] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.147863] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.148050] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 806.148138] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87b3ce5d-4e66-4ca5-b029-c34b02c71e1e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.150075] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21b8651f-f54e-4b2a-b7c1-710a8f3390b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.155787] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 806.155787] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ed8834-e76a-acc3-1917-9368b022af38" [ 806.155787] env[70013]: _type = "Task" [ 806.155787] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.164017] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.164017] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.164017] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1b7ac7-9840-4e28-bc58-b1751b9f7183 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.176329] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ed8834-e76a-acc3-1917-9368b022af38, 'name': SearchDatastore_Task, 'duration_secs': 0.011444} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.176846] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 806.176846] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52da49f3-c24b-b29d-8ce5-7d8e1b741d78" [ 806.176846] env[70013]: _type = "Task" [ 806.176846] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.177040] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.177207] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.177430] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.189021] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52da49f3-c24b-b29d-8ce5-7d8e1b741d78, 'name': SearchDatastore_Task, 'duration_secs': 0.011159} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.189021] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67b3855b-cd94-446a-9b06-eb87fe214773 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.194904] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 806.194904] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5215f39d-d005-87a3-676f-1aa1a3f90335" [ 806.194904] env[70013]: _type = "Task" [ 806.194904] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.209387] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5215f39d-d005-87a3-676f-1aa1a3f90335, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.329065] env[70013]: DEBUG nova.compute.utils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 806.334316] env[70013]: DEBUG nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 806.334316] env[70013]: DEBUG nova.network.neutron [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.477470] env[70013]: DEBUG oslo_concurrency.lockutils [req-239532c5-5df3-4e12-8c48-b6afcedaa5b7 req-610b61ab-835a-4ab4-a4ba-dc6957eda4bd service nova] Releasing lock "refresh_cache-77c12460-9cfa-41c9-a210-238a470d9ccd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.523734] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230796, 'name': Rename_Task, 'duration_secs': 0.17942} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.523896] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.524176] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd893bb2-3f6d-48bc-9333-3ef228ff6e77 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.532489] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 806.532489] env[70013]: value = "task-4230797" [ 806.532489] env[70013]: _type = "Task" [ 806.532489] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.543756] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230797, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.677373] env[70013]: DEBUG nova.policy [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '78c46c65317845ff9cfc895832fafaa9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1fa3c6a4122484c831621d0f108a5f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 806.709074] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5215f39d-d005-87a3-676f-1aa1a3f90335, 'name': SearchDatastore_Task, 'duration_secs': 0.011612} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.710413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.710413] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 77c12460-9cfa-41c9-a210-238a470d9ccd/77c12460-9cfa-41c9-a210-238a470d9ccd.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 806.710413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.710829] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.711135] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-811080c3-433b-4cd6-89ad-b8d66c4f03b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.713256] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59dfefda-0bb2-4129-8588-6781f41181aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.721788] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 806.721788] env[70013]: value = "task-4230798" [ 806.721788] env[70013]: _type = "Task" [ 806.721788] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.722880] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.723067] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.727194] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c268b4a3-88b0-46f3-b9b5-efdf327b2d9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.737536] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.742520] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 806.742520] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521ef9c2-5e8b-1abb-477c-3a2ff8ea4ba1" [ 806.742520] env[70013]: _type = "Task" [ 806.742520] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.752966] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521ef9c2-5e8b-1abb-477c-3a2ff8ea4ba1, 'name': SearchDatastore_Task, 'duration_secs': 0.010831} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.753554] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-386afbe7-aafd-42ab-9458-2df94f67460f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.761839] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 806.761839] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5219bee7-421b-d21f-2c0b-a96097a0aa09" [ 806.761839] env[70013]: _type = "Task" [ 806.761839] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.771366] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5219bee7-421b-d21f-2c0b-a96097a0aa09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.836634] env[70013]: DEBUG nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 807.032279] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c8e19f-ef32-4584-bde2-fcb2b873e2d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.050647] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230797, 'name': PowerOnVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.054268] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6784205f-0bd2-4f6b-9fda-1c34b7923d07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.105305] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fbbd8f-ab19-4a5f-abac-e4e858709a09 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.118888] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50aa2f4-2aee-42be-b54b-1b7f47cc490e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.141512] env[70013]: DEBUG nova.compute.provider_tree [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.244243] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230798, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.276493] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5219bee7-421b-d21f-2c0b-a96097a0aa09, 'name': SearchDatastore_Task, 'duration_secs': 0.010576} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.276493] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.276493] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] ac8aa41c-375a-4231-9a5d-12d6c4bba44c/ac8aa41c-375a-4231-9a5d-12d6c4bba44c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 807.276493] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6949b91a-e550-42f9-b4e0-bd3dc9f9bbff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.286067] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 807.286067] env[70013]: value = "task-4230799" [ 807.286067] env[70013]: _type = "Task" [ 807.286067] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.296035] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230799, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.310257] env[70013]: DEBUG nova.network.neutron [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Successfully created port: 2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.345276] env[70013]: INFO nova.virt.block_device [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Booting with volume 19da385b-04e0-4e04-9b4d-39b9370bec62 at /dev/sda [ 807.403858] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2508825-b64d-4652-9f72-44e47f42a433 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.416017] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262ed5b6-7efe-43ff-bfbe-282479d86740 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.453821] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f0744fa-79d4-42e7-9a47-718f3fe3c1ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.462043] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0615a442-3315-4988-b031-d5e3fd259b17 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.507601] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d7eb91-7215-4c13-ae8b-8f0c07fe0f2a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.522500] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2239e12d-d0ac-4b40-9826-f8b90ce5d024 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.542241] env[70013]: DEBUG nova.virt.block_device [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updating existing volume attachment record: ca9bfe48-7cf2-4dc5-8beb-16727e4ee1bb {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 807.562692] env[70013]: DEBUG oslo_vmware.api [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230797, 'name': PowerOnVM_Task, 'duration_secs': 0.603292} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.562982] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.563201] env[70013]: INFO nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Took 7.93 seconds to spawn the instance on the hypervisor. [ 807.563376] env[70013]: DEBUG nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 807.565752] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d35737-7210-4263-86d5-2e8862c5f86f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.645876] env[70013]: DEBUG nova.scheduler.client.report [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 807.741659] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230798, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564839} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.741953] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 77c12460-9cfa-41c9-a210-238a470d9ccd/77c12460-9cfa-41c9-a210-238a470d9ccd.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.742247] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.742454] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac460802-0bc9-49f9-bba6-8d1ec3620050 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.754402] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 807.754402] env[70013]: value = "task-4230800" [ 807.754402] env[70013]: _type = "Task" [ 807.754402] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.770896] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230800, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.794547] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230799, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.097320] env[70013]: INFO nova.compute.manager [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Took 42.28 seconds to build instance. [ 808.156089] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.335s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.158768] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.121s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.159015] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.159761] env[70013]: INFO nova.compute.manager [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Successfully reverted task state from None on failure for instance. [ 808.163029] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.167020] env[70013]: INFO nova.compute.claims [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server [None req-c8419cb7-66bb-4c2c-a2d6-fdf8fe07dfbe tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server yield [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-16205499-7cc0-4a6e-a789-bd1751daf972"}]} [ 808.169645] env[70013]: ERROR oslo_messaging.rpc.server [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 808.169963] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 808.170583] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 808.171070] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 808.171487] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 808.172107] env[70013]: ERROR oslo_messaging.rpc.server [ 808.177116] env[70013]: INFO nova.scheduler.client.report [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Deleted allocations for instance 062eff58-95fe-4c9b-a586-7e7434c77adf [ 808.266412] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230800, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086301} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.266699] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.268174] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a955751-a45d-47a5-a0a4-f0deb38dbe77 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.292856] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 77c12460-9cfa-41c9-a210-238a470d9ccd/77c12460-9cfa-41c9-a210-238a470d9ccd.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.293308] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d9f5d59-0936-45fc-a3a9-04cc657a3ec1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.320191] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230799, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518562} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.321798] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] ac8aa41c-375a-4231-9a5d-12d6c4bba44c/ac8aa41c-375a-4231-9a5d-12d6c4bba44c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 808.322068] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.322751] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 808.322751] env[70013]: value = "task-4230801" [ 808.322751] env[70013]: _type = "Task" [ 808.322751] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.322751] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37d01647-b933-4407-ba4f-07008ea18582 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.333609] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230801, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.335578] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 808.335578] env[70013]: value = "task-4230802" [ 808.335578] env[70013]: _type = "Task" [ 808.335578] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.346818] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230802, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.600474] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90186932-37c8-47a6-a0b8-86f4b485265e tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.046s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.688029] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52120071-e6fb-4293-adef-7738fed84ba9 tempest-ServerDiagnosticsV248Test-777977431 tempest-ServerDiagnosticsV248Test-777977431-project-member] Lock "062eff58-95fe-4c9b-a586-7e7434c77adf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.222s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.747534] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 808.747813] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.837657] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230801, 'name': ReconfigVM_Task, 'duration_secs': 0.288857} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.842701] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 77c12460-9cfa-41c9-a210-238a470d9ccd/77c12460-9cfa-41c9-a210-238a470d9ccd.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.843382] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-074a948f-e6ac-4010-9bdd-c0152a14ebd4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.851787] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230802, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074032} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.853362] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.853842] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 808.853842] env[70013]: value = "task-4230803" [ 808.853842] env[70013]: _type = "Task" [ 808.853842] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.854934] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e01601-2bab-4092-9d55-92025adebee5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.881306] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] ac8aa41c-375a-4231-9a5d-12d6c4bba44c/ac8aa41c-375a-4231-9a5d-12d6c4bba44c.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.885382] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cab529c3-bca3-4d3d-959c-5a4aa1a7288c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.900966] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230803, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.907437] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 808.907437] env[70013]: value = "task-4230804" [ 808.907437] env[70013]: _type = "Task" [ 808.907437] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.917072] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230804, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.105576] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 809.327119] env[70013]: DEBUG nova.compute.manager [req-baecd7ed-d0d8-489c-89a4-7ce205fb3f74 req-201827d4-9382-456d-81fc-c80408dcc4af service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Received event network-vif-plugged-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 809.327374] env[70013]: DEBUG oslo_concurrency.lockutils [req-baecd7ed-d0d8-489c-89a4-7ce205fb3f74 req-201827d4-9382-456d-81fc-c80408dcc4af service nova] Acquiring lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.329197] env[70013]: DEBUG oslo_concurrency.lockutils [req-baecd7ed-d0d8-489c-89a4-7ce205fb3f74 req-201827d4-9382-456d-81fc-c80408dcc4af service nova] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.329197] env[70013]: DEBUG oslo_concurrency.lockutils [req-baecd7ed-d0d8-489c-89a4-7ce205fb3f74 req-201827d4-9382-456d-81fc-c80408dcc4af service nova] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.329197] env[70013]: DEBUG nova.compute.manager [req-baecd7ed-d0d8-489c-89a4-7ce205fb3f74 req-201827d4-9382-456d-81fc-c80408dcc4af service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] No waiting events found dispatching network-vif-plugged-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 809.329197] env[70013]: WARNING nova.compute.manager [req-baecd7ed-d0d8-489c-89a4-7ce205fb3f74 req-201827d4-9382-456d-81fc-c80408dcc4af service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Received unexpected event network-vif-plugged-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 for instance with vm_state building and task_state spawning. [ 809.377541] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230803, 'name': Rename_Task, 'duration_secs': 0.15373} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.380306] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.380306] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8aefd9bb-4a11-419e-8380-ccbcb9ca9628 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.390128] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 809.390128] env[70013]: value = "task-4230805" [ 809.390128] env[70013]: _type = "Task" [ 809.390128] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.414205] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.415398] env[70013]: DEBUG nova.network.neutron [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Successfully updated port: 2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.429165] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230804, 'name': ReconfigVM_Task, 'duration_secs': 0.313803} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.432397] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Reconfigured VM instance instance-00000020 to attach disk [datastore2] ac8aa41c-375a-4231-9a5d-12d6c4bba44c/ac8aa41c-375a-4231-9a5d-12d6c4bba44c.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.433987] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3cb611db-a70f-42f1-90fb-7af43dfb2410 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.443294] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 809.443294] env[70013]: value = "task-4230806" [ 809.443294] env[70013]: _type = "Task" [ 809.443294] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.456673] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230806, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.630445] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.688686] env[70013]: DEBUG nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 809.689394] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 809.689650] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.689816] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 809.690020] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.690187] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 809.690345] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 809.690565] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 809.690749] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 809.691256] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 809.691504] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 809.691879] env[70013]: DEBUG nova.virt.hardware [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 809.695544] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2349e8f3-bf4f-4b0d-9c20-4add1e2a0176 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.706084] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77be8b6a-1e67-419a-a31e-323182126bab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.784941] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c997c7-1952-411b-8c0f-230cfd18ceac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.794141] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4e3e0d-9ab2-45c5-8825-485c15e30c79 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.797792] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.798977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.798977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.798977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.798977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.830829] env[70013]: INFO nova.compute.manager [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Terminating instance [ 809.833412] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e115f4dd-ba69-4ad0-93c5-27f35cb02a12 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.844998] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fc4ef9-df99-4693-b2f8-0360b8aeaf91 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.863293] env[70013]: DEBUG nova.compute.provider_tree [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 809.901608] env[70013]: DEBUG oslo_vmware.api [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230805, 'name': PowerOnVM_Task, 'duration_secs': 0.475684} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.901881] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 809.902127] env[70013]: INFO nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Took 7.58 seconds to spawn the instance on the hypervisor. [ 809.902318] env[70013]: DEBUG nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 809.903159] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b80fba6-8375-4623-8bbb-a1be68f75de8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.922720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquiring lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.922720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquired lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.922720] env[70013]: DEBUG nova.network.neutron [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.952720] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230806, 'name': Rename_Task, 'duration_secs': 0.148312} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.953695] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.953955] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27fcad9f-13d5-496f-bdcc-dde4e2a1a698 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.962124] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 809.962124] env[70013]: value = "task-4230807" [ 809.962124] env[70013]: _type = "Task" [ 809.962124] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.971543] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230807, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.344559] env[70013]: DEBUG nova.compute.manager [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 810.344559] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.345232] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d32410e-6168-4868-b4d8-cdff87948d11 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.354038] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.354354] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e2bee40-ad88-4054-8610-a663dbce7f5d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.363390] env[70013]: DEBUG oslo_vmware.api [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 810.363390] env[70013]: value = "task-4230808" [ 810.363390] env[70013]: _type = "Task" [ 810.363390] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.378370] env[70013]: DEBUG oslo_vmware.api [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230808, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.386891] env[70013]: ERROR nova.scheduler.client.report [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [req-0573c22d-b878-41cd-a705-6ee663907783] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0573c22d-b878-41cd-a705-6ee663907783"}]} [ 810.387412] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.387933] env[70013]: ERROR nova.compute.manager [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 810.387933] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Traceback (most recent call last): [ 810.387933] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 810.387933] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] yield [ 810.387933] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 810.387933] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] self.set_inventory_for_provider( [ 810.387933] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 810.387933] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0573c22d-b878-41cd-a705-6ee663907783"}]} [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] During handling of the above exception, another exception occurred: [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Traceback (most recent call last): [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] with self.rt.instance_claim(context, instance, node, allocs, [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 810.388202] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] return f(*args, **kwargs) [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] self._update(elevated, cn) [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] self._update_to_placement(context, compute_node, startup) [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] return attempt.get(self._wrap_exception) [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] six.reraise(self.value[0], self.value[1], self.value[2]) [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] raise value [ 810.388512] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] self.reportclient.update_from_provider_tree( [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] with catch_all(pd.uuid): [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] self.gen.throw(typ, value, traceback) [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] raise exception.ResourceProviderSyncFailed() [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 810.388887] env[70013]: ERROR nova.compute.manager [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] [ 810.389224] env[70013]: DEBUG nova.compute.utils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 810.389869] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.613s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.390065] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.390225] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 810.390510] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.650s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.390710] env[70013]: DEBUG nova.objects.instance [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lazy-loading 'resources' on Instance uuid 85488f5e-bd8a-4bcf-8420-744c54176c2e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 810.392580] env[70013]: DEBUG nova.compute.manager [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Build of instance e39ec7f2-76d7-4794-9812-e4eb23a405e7 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 810.392580] env[70013]: DEBUG nova.compute.manager [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 810.393654] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "refresh_cache-e39ec7f2-76d7-4794-9812-e4eb23a405e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.393654] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquired lock "refresh_cache-e39ec7f2-76d7-4794-9812-e4eb23a405e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.393654] env[70013]: DEBUG nova.network.neutron [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 810.395425] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca111234-9f0b-4cb6-a5e1-41e9b2dc2ff1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.406133] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9ef597-5d5e-4b75-86a1-5a48ca44c1e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.431306] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbec12a-baec-4ced-8c24-fd0b03753ede {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.435220] env[70013]: INFO nova.compute.manager [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Took 43.53 seconds to build instance. [ 810.449403] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f5a2d1-8b0a-4884-a82f-c5676e9faec0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.484851] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180343MB free_disk=0GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 810.486038] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.496034] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230807, 'name': PowerOnVM_Task} progress is 96%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.496925] env[70013]: DEBUG nova.network.neutron [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.728614] env[70013]: DEBUG nova.network.neutron [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updating instance_info_cache with network_info: [{"id": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "address": "fa:16:3e:9b:0a:0d", "network": {"id": "b250add1-805e-4928-b1de-496c88097a4e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-15368811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1fa3c6a4122484c831621d0f108a5f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f01bbee7-8b9a-46be-891e-59d8142fb359", "external-id": "nsx-vlan-transportzone-145", "segmentation_id": 145, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd50eea-d7", "ovs_interfaceid": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.878351] env[70013]: DEBUG oslo_vmware.api [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230808, 'name': PowerOffVM_Task, 'duration_secs': 0.315935} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.878904] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.879431] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.879872] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f25ae25-126d-4fb4-80c7-4f36d1dcd278 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.914465] env[70013]: DEBUG nova.scheduler.client.report [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 810.924462] env[70013]: DEBUG nova.network.neutron [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.932483] env[70013]: DEBUG nova.scheduler.client.report [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 810.932782] env[70013]: DEBUG nova.compute.provider_tree [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.938443] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1e2c263-22cd-4ca1-9956-1a952cac3598 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.434s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.948745] env[70013]: DEBUG nova.scheduler.client.report [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 810.956923] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.956923] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.956923] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Deleting the datastore file [datastore2] 2ee62ad3-0125-47dc-b163-7d15b6a17c8a {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.960248] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5618da0b-e009-4d4e-ad17-52a0361e3eb3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.974413] env[70013]: DEBUG oslo_vmware.api [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for the task: (returnval){ [ 810.974413] env[70013]: value = "task-4230810" [ 810.974413] env[70013]: _type = "Task" [ 810.974413] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.977389] env[70013]: DEBUG nova.scheduler.client.report [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 810.989688] env[70013]: DEBUG oslo_vmware.api [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.999494] env[70013]: DEBUG oslo_vmware.api [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230807, 'name': PowerOnVM_Task, 'duration_secs': 0.698743} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.999848] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.000046] env[70013]: INFO nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Took 5.90 seconds to spawn the instance on the hypervisor. [ 811.000291] env[70013]: DEBUG nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 811.001180] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6628485-c76e-48af-b62f-382e113d5084 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.025773] env[70013]: DEBUG nova.network.neutron [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.230897] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Releasing lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.231303] env[70013]: DEBUG nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance network_info: |[{"id": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "address": "fa:16:3e:9b:0a:0d", "network": {"id": "b250add1-805e-4928-b1de-496c88097a4e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-15368811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1fa3c6a4122484c831621d0f108a5f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f01bbee7-8b9a-46be-891e-59d8142fb359", "external-id": "nsx-vlan-transportzone-145", "segmentation_id": 145, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd50eea-d7", "ovs_interfaceid": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 811.233840] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:0a:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f01bbee7-8b9a-46be-891e-59d8142fb359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.241882] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Creating folder: Project (a1fa3c6a4122484c831621d0f108a5f6). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.245145] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2082e6b9-0941-417e-aac0-00068c78f166 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.261332] env[70013]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 811.261510] env[70013]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=70013) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 811.262264] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Folder already exists: Project (a1fa3c6a4122484c831621d0f108a5f6). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 811.262537] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Creating folder: Instances. Parent ref: group-v837053. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.262833] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76df433d-c545-4144-9192-98756bae1099 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.279181] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Created folder: Instances in parent group-v837053. [ 811.279608] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 811.279865] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.280148] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c7f9d1e-681a-4453-9cfc-e78a93e9bf43 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.306379] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.306379] env[70013]: value = "task-4230813" [ 811.306379] env[70013]: _type = "Task" [ 811.306379] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.320439] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230813, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.400856] env[70013]: DEBUG nova.compute.manager [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Received event network-changed-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 811.401178] env[70013]: DEBUG nova.compute.manager [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Refreshing instance network info cache due to event network-changed-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 811.401645] env[70013]: DEBUG oslo_concurrency.lockutils [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] Acquiring lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.401743] env[70013]: DEBUG oslo_concurrency.lockutils [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] Acquired lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.401992] env[70013]: DEBUG nova.network.neutron [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Refreshing network info cache for port 2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.441957] env[70013]: DEBUG nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 811.485031] env[70013]: DEBUG oslo_vmware.api [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Task: {'id': task-4230810, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.473975} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.485185] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.485362] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 811.485548] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.485735] env[70013]: INFO nova.compute.manager [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 811.485976] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 811.486362] env[70013]: DEBUG nova.compute.manager [-] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 811.486460] env[70013]: DEBUG nova.network.neutron [-] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 811.526839] env[70013]: INFO nova.compute.manager [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Took 38.64 seconds to build instance. [ 811.532261] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Releasing lock "refresh_cache-e39ec7f2-76d7-4794-9812-e4eb23a405e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.532393] env[70013]: DEBUG nova.compute.manager [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 811.532583] env[70013]: DEBUG nova.compute.manager [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 811.532796] env[70013]: DEBUG nova.network.neutron [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 811.554826] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eace0111-e1b6-4e91-9a96-0f2c3ecb530c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.558273] env[70013]: DEBUG nova.network.neutron [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.569293] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2085a6-92b7-4ce2-883f-80a6ea50938e tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Suspending the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 811.569861] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-862c585b-92a2-461f-ad1a-5af56537f997 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.579895] env[70013]: DEBUG oslo_vmware.api [None req-0a2085a6-92b7-4ce2-883f-80a6ea50938e tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 811.579895] env[70013]: value = "task-4230814" [ 811.579895] env[70013]: _type = "Task" [ 811.579895] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.594232] env[70013]: DEBUG oslo_vmware.api [None req-0a2085a6-92b7-4ce2-883f-80a6ea50938e tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230814, 'name': SuspendVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.650121] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce4381c-dc26-45fe-912a-7169c2b71869 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.660576] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5ded52-0769-4f23-a700-34429348d87f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.697641] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4990e9-ef7d-474f-a63e-73a5fe217740 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.705863] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3532963e-d7d8-4540-bac9-35e6d1f72fd9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.729638] env[70013]: DEBUG nova.compute.provider_tree [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.823188] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230813, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.986092] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.029775] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d26ed843-e8bd-44ef-a929-59f819bddfe3 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.083s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.065263] env[70013]: DEBUG nova.network.neutron [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.092611] env[70013]: DEBUG oslo_vmware.api [None req-0a2085a6-92b7-4ce2-883f-80a6ea50938e tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230814, 'name': SuspendVM_Task} progress is 62%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.254573] env[70013]: ERROR nova.scheduler.client.report [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [req-b71a728c-655c-4394-aca1-0ecf06dcab40] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b71a728c-655c-4394-aca1-0ecf06dcab40"}]} [ 812.255664] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.256088] env[70013]: ERROR nova.compute.manager [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 812.256088] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Traceback (most recent call last): [ 812.256088] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 812.256088] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] yield [ 812.256088] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 812.256088] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self.set_inventory_for_provider( [ 812.256088] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 812.256088] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b71a728c-655c-4394-aca1-0ecf06dcab40"}]} [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] During handling of the above exception, another exception occurred: [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Traceback (most recent call last): [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self._delete_instance(context, instance, bdms) [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 812.256330] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self._complete_deletion(context, instance) [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self._update_resource_tracker(context, instance) [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self.rt.update_usage(context, instance, instance.node) [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] return f(*args, **kwargs) [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self._update(context.elevated(), self.compute_nodes[nodename]) [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self._update_to_placement(context, compute_node, startup) [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 812.256672] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] return attempt.get(self._wrap_exception) [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] six.reraise(self.value[0], self.value[1], self.value[2]) [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] raise value [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self.reportclient.update_from_provider_tree( [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] with catch_all(pd.uuid): [ 812.257056] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 812.257464] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] self.gen.throw(typ, value, traceback) [ 812.257464] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 812.257464] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] raise exception.ResourceProviderSyncFailed() [ 812.257464] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 812.257464] env[70013]: ERROR nova.compute.manager [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] [ 812.259098] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.222s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.260732] env[70013]: INFO nova.compute.claims [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.318990] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230813, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.336461] env[70013]: DEBUG nova.network.neutron [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updated VIF entry in instance network info cache for port 2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.337326] env[70013]: DEBUG nova.network.neutron [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updating instance_info_cache with network_info: [{"id": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "address": "fa:16:3e:9b:0a:0d", "network": {"id": "b250add1-805e-4928-b1de-496c88097a4e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-15368811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1fa3c6a4122484c831621d0f108a5f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f01bbee7-8b9a-46be-891e-59d8142fb359", "external-id": "nsx-vlan-transportzone-145", "segmentation_id": 145, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd50eea-d7", "ovs_interfaceid": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.367832] env[70013]: DEBUG nova.network.neutron [-] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.532628] env[70013]: DEBUG nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 812.569568] env[70013]: INFO nova.compute.manager [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: e39ec7f2-76d7-4794-9812-e4eb23a405e7] Took 1.04 seconds to deallocate network for instance. [ 812.591177] env[70013]: DEBUG oslo_vmware.api [None req-0a2085a6-92b7-4ce2-883f-80a6ea50938e tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230814, 'name': SuspendVM_Task, 'duration_secs': 0.674904} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.591474] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2085a6-92b7-4ce2-883f-80a6ea50938e tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Suspended the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 812.591685] env[70013]: DEBUG nova.compute.manager [None req-0a2085a6-92b7-4ce2-883f-80a6ea50938e tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 812.592544] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3249f33c-c0ce-4248-a841-be851ce3ccc8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.768149] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.547s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.818112] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230813, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.845025] env[70013]: DEBUG oslo_concurrency.lockutils [req-875f6203-4689-4d2c-bd80-e7315062f21b req-5b9e90c0-3a37-407f-9434-eec9ceef5640 service nova] Releasing lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 812.870122] env[70013]: INFO nova.compute.manager [-] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Took 1.38 seconds to deallocate network for instance. [ 813.057349] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.123555] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.124678] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.293443] env[70013]: DEBUG nova.scheduler.client.report [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 813.308787] env[70013]: DEBUG nova.scheduler.client.report [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 813.311023] env[70013]: DEBUG nova.compute.provider_tree [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.325820] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230813, 'name': CreateVM_Task, 'duration_secs': 1.514532} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.327434] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 813.328744] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'ca9bfe48-7cf2-4dc5-8beb-16727e4ee1bb', 'guest_format': None, 'device_type': None, 'mount_device': '/dev/sda', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837062', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'name': 'volume-19da385b-04e0-4e04-9b4d-39b9370bec62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5c5fbdae-4916-4beb-ada0-57f36fb0f84c', 'attached_at': '', 'detached_at': '', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'serial': '19da385b-04e0-4e04-9b4d-39b9370bec62'}, 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=70013) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 813.329172] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Root volume attach. Driver type: vmdk {{(pid=70013) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 813.332368] env[70013]: DEBUG nova.scheduler.client.report [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 813.335044] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d67602-4db8-47b1-941e-0fa204099b5f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.346133] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad3482b-26c9-4c09-bfe2-3981adf6b7cd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.356021] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69908e08-c19a-499a-a529-b01022cda384 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.360872] env[70013]: DEBUG nova.scheduler.client.report [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 813.362983] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-93247dab-e8a0-49f0-aab1-8f41d6cc4cc4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.371939] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 813.371939] env[70013]: value = "task-4230815" [ 813.371939] env[70013]: _type = "Task" [ 813.371939] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.380940] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.387124] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230815, 'name': RelocateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.460130] env[70013]: DEBUG nova.compute.manager [req-46423473-2ea9-485b-b051-e72398b677ac req-66dd5245-646b-48ed-b989-09229f2b2071 service nova] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Received event network-vif-deleted-c7fc6ffa-a891-4805-adfe-0a1ed770b245 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 813.540671] env[70013]: DEBUG nova.compute.manager [None req-813d5e50-359a-42a4-9e3b-4af925f2946b tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 813.541747] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48fd72f-fbb8-4bc1-a4ed-b6b793e3f311 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.613488] env[70013]: INFO nova.scheduler.client.report [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Deleted allocations for instance e39ec7f2-76d7-4794-9812-e4eb23a405e7 [ 813.887341] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230815, 'name': RelocateVM_Task, 'duration_secs': 0.42519} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.887341] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 813.887341] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837062', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'name': 'volume-19da385b-04e0-4e04-9b4d-39b9370bec62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5c5fbdae-4916-4beb-ada0-57f36fb0f84c', 'attached_at': '', 'detached_at': '', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'serial': '19da385b-04e0-4e04-9b4d-39b9370bec62'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 813.887731] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd37408-032d-49e8-87ee-9faee8b208c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.907034] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce94c58-25d9-496d-8cc9-7a24cf84ee64 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.929158] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] volume-19da385b-04e0-4e04-9b4d-39b9370bec62/volume-19da385b-04e0-4e04-9b4d-39b9370bec62.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.932326] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.932569] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.932772] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.932953] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.933132] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.934824] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00295b7e-8024-4b79-839f-65e0b2f129eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.949406] env[70013]: INFO nova.compute.manager [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Terminating instance [ 813.957835] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 813.957835] env[70013]: value = "task-4230816" [ 813.957835] env[70013]: _type = "Task" [ 813.957835] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.968480] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.054755] env[70013]: INFO nova.compute.manager [None req-813d5e50-359a-42a4-9e3b-4af925f2946b tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] instance snapshotting [ 814.055416] env[70013]: DEBUG nova.objects.instance [None req-813d5e50-359a-42a4-9e3b-4af925f2946b tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lazy-loading 'flavor' on Instance uuid ac8aa41c-375a-4231-9a5d-12d6c4bba44c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.123648] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4e965270-43f5-4641-b219-5ae768413956 tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "e39ec7f2-76d7-4794-9812-e4eb23a405e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.574s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.139393] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63dcccd-6140-4789-a3c9-cc47cadc28d8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.148791] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f66a41-b4e8-4702-aaa4-dabd03109bba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.183809] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82206686-9de9-4538-9d52-6df2ce9df8a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.193141] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ea48cd-f427-4358-a9c6-0acaee2b478d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.209962] env[70013]: DEBUG nova.compute.provider_tree [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.285833] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.455638] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "refresh_cache-ac8aa41c-375a-4231-9a5d-12d6c4bba44c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.455841] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquired lock "refresh_cache-ac8aa41c-375a-4231-9a5d-12d6c4bba44c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.456128] env[70013]: DEBUG nova.network.neutron [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.469055] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230816, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.567837] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff17ffd-dc62-46ef-b24b-f22a9075ee7b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.587406] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a5ea16-e585-49b2-8945-81dea6847ebe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.629048] env[70013]: DEBUG nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 814.734618] env[70013]: ERROR nova.scheduler.client.report [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [req-494e51e0-82cb-4a07-af10-13e6ac5ae87e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-494e51e0-82cb-4a07-af10-13e6ac5ae87e"}]} [ 814.735289] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.737426] env[70013]: ERROR nova.compute.manager [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 814.737426] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] Traceback (most recent call last): [ 814.737426] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 814.737426] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] yield [ 814.737426] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 814.737426] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] self.set_inventory_for_provider( [ 814.737426] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 814.737426] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-494e51e0-82cb-4a07-af10-13e6ac5ae87e"}]} [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] During handling of the above exception, another exception occurred: [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] Traceback (most recent call last): [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] with self.rt.instance_claim(context, instance, node, allocs, [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 814.737718] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] return f(*args, **kwargs) [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] self._update(elevated, cn) [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] self._update_to_placement(context, compute_node, startup) [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] return attempt.get(self._wrap_exception) [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] raise value [ 814.737958] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] self.reportclient.update_from_provider_tree( [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] with catch_all(pd.uuid): [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] self.gen.throw(typ, value, traceback) [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] raise exception.ResourceProviderSyncFailed() [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 814.738273] env[70013]: ERROR nova.compute.manager [instance: f1871673-e764-460b-adce-1742e7c105c6] [ 814.738528] env[70013]: DEBUG nova.compute.utils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 814.738528] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.536s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.739528] env[70013]: INFO nova.compute.claims [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.744282] env[70013]: DEBUG nova.compute.manager [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Build of instance f1871673-e764-460b-adce-1742e7c105c6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 814.744591] env[70013]: DEBUG nova.compute.manager [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 814.744834] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "refresh_cache-f1871673-e764-460b-adce-1742e7c105c6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.744979] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquired lock "refresh_cache-f1871673-e764-460b-adce-1742e7c105c6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.745159] env[70013]: DEBUG nova.network.neutron [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.975149] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230816, 'name': ReconfigVM_Task, 'duration_secs': 0.709253} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.975149] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Reconfigured VM instance instance-00000021 to attach disk [datastore1] volume-19da385b-04e0-4e04-9b4d-39b9370bec62/volume-19da385b-04e0-4e04-9b4d-39b9370bec62.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 814.980164] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9fd54a5-a681-4857-8e02-e3a5bd5d70c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.991630] env[70013]: DEBUG nova.network.neutron [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.999148] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 814.999148] env[70013]: value = "task-4230817" [ 814.999148] env[70013]: _type = "Task" [ 814.999148] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.007927] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230817, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.058336] env[70013]: DEBUG nova.network.neutron [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.099545] env[70013]: DEBUG nova.compute.manager [None req-813d5e50-359a-42a4-9e3b-4af925f2946b tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Instance disappeared during snapshot {{(pid=70013) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4595}} [ 815.161211] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.257875] env[70013]: DEBUG nova.compute.manager [None req-813d5e50-359a-42a4-9e3b-4af925f2946b tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Found 0 images (rotation: 2) {{(pid=70013) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 815.271298] env[70013]: DEBUG nova.compute.manager [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 815.274081] env[70013]: DEBUG nova.network.neutron [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.277074] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317adcda-5c73-41c3-bd18-1208d8a9293a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.509801] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230817, 'name': ReconfigVM_Task, 'duration_secs': 0.170489} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.510412] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837062', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'name': 'volume-19da385b-04e0-4e04-9b4d-39b9370bec62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5c5fbdae-4916-4beb-ada0-57f36fb0f84c', 'attached_at': '', 'detached_at': '', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'serial': '19da385b-04e0-4e04-9b4d-39b9370bec62'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 815.511024] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf02f060-c6a4-417d-be7d-619b2ee7aeb9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.518582] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 815.518582] env[70013]: value = "task-4230818" [ 815.518582] env[70013]: _type = "Task" [ 815.518582] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.531336] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230818, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.562728] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Releasing lock "refresh_cache-ac8aa41c-375a-4231-9a5d-12d6c4bba44c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 815.563183] env[70013]: DEBUG nova.compute.manager [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 815.563384] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.564520] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2bc7308-4f23-4d7d-b136-1582edab65d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.572600] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.572968] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86da132a-df00-43a6-8e20-6d73f806ed28 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.581251] env[70013]: DEBUG oslo_vmware.api [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 815.581251] env[70013]: value = "task-4230819" [ 815.581251] env[70013]: _type = "Task" [ 815.581251] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.597765] env[70013]: DEBUG oslo_vmware.api [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230819, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.599194] env[70013]: DEBUG nova.network.neutron [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.782175] env[70013]: DEBUG nova.scheduler.client.report [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 815.789278] env[70013]: INFO nova.compute.manager [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] instance snapshotting [ 815.789527] env[70013]: WARNING nova.compute.manager [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 815.793082] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cedae4-a816-4ffd-b3d4-dbd63c8377fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.797796] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "3e938bb3-54a6-49f8-809d-d5aee0349eae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.798417] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.798649] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "3e938bb3-54a6-49f8-809d-d5aee0349eae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.798834] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.799014] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.814586] env[70013]: INFO nova.compute.manager [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Terminating instance [ 815.816935] env[70013]: DEBUG nova.scheduler.client.report [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 815.817175] env[70013]: DEBUG nova.compute.provider_tree [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 815.819774] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89f4e31-cf92-49ac-9c68-6e05bdae34ea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.836230] env[70013]: DEBUG nova.scheduler.client.report [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 815.856533] env[70013]: DEBUG nova.scheduler.client.report [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 816.031093] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230818, 'name': Rename_Task, 'duration_secs': 0.157595} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.031684] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.031985] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-727bd4bb-61a3-4c2f-94bc-244420d9c64b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.040627] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 816.040627] env[70013]: value = "task-4230820" [ 816.040627] env[70013]: _type = "Task" [ 816.040627] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.049317] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230820, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.093227] env[70013]: DEBUG oslo_vmware.api [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230819, 'name': PowerOffVM_Task, 'duration_secs': 0.13139} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.096025] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.096144] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.096598] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61217fb2-e32b-4eb1-b473-cc3b985e1de0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.101895] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Releasing lock "refresh_cache-f1871673-e764-460b-adce-1742e7c105c6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 816.102138] env[70013]: DEBUG nova.compute.manager [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 816.102308] env[70013]: DEBUG nova.compute.manager [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 816.102484] env[70013]: DEBUG nova.network.neutron [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.123253] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.123386] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.123529] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Deleting the datastore file [datastore2] ac8aa41c-375a-4231-9a5d-12d6c4bba44c {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.124212] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbfe425e-6eee-4eb0-95be-cc7d20bd9941 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.131492] env[70013]: DEBUG oslo_vmware.api [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for the task: (returnval){ [ 816.131492] env[70013]: value = "task-4230822" [ 816.131492] env[70013]: _type = "Task" [ 816.131492] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.146281] env[70013]: DEBUG nova.network.neutron [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.147619] env[70013]: DEBUG oslo_vmware.api [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.325538] env[70013]: DEBUG nova.compute.manager [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 816.325766] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.326676] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156580c4-99b1-4dd5-a2ea-f7b8bbdbda41 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.332635] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 816.333773] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2baa9ab3-4d08-4a76-b1f3-316810d1c9a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.338704] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.339071] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55dba7ef-48ec-4c67-8f75-7796287a836a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.345833] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 816.345833] env[70013]: value = "task-4230823" [ 816.345833] env[70013]: _type = "Task" [ 816.345833] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.347248] env[70013]: DEBUG oslo_vmware.api [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 816.347248] env[70013]: value = "task-4230824" [ 816.347248] env[70013]: _type = "Task" [ 816.347248] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.361871] env[70013]: DEBUG oslo_vmware.api [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.365054] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230823, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.373333] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33b6629-91e9-4cfa-8257-ad3b4a3c90cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.381656] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9f82af-1045-4c5e-b040-66805d31f80d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.417314] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02e41dd-d611-4494-9e90-a17be3ac086f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.428268] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e710e267-6cde-4309-8ace-9908d2cad851 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.446505] env[70013]: DEBUG nova.compute.provider_tree [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 816.551680] env[70013]: DEBUG oslo_vmware.api [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230820, 'name': PowerOnVM_Task, 'duration_secs': 0.474269} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.552056] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.552292] env[70013]: INFO nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Took 6.86 seconds to spawn the instance on the hypervisor. [ 816.552549] env[70013]: DEBUG nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 816.553421] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3f8131-6020-47ec-ad76-f6866d1050cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.643875] env[70013]: DEBUG oslo_vmware.api [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Task: {'id': task-4230822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.117866} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.644274] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.644381] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.644515] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.644747] env[70013]: INFO nova.compute.manager [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Took 1.08 seconds to destroy the instance on the hypervisor. [ 816.644997] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 816.645216] env[70013]: DEBUG nova.compute.manager [-] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 816.645314] env[70013]: DEBUG nova.network.neutron [-] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.649023] env[70013]: DEBUG nova.network.neutron [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.667939] env[70013]: DEBUG nova.network.neutron [-] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.864862] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230823, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.868519] env[70013]: DEBUG oslo_vmware.api [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230824, 'name': PowerOffVM_Task, 'duration_secs': 0.212285} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.868820] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.868996] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.869339] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-069f6f19-0f73-40b2-b6e2-a9774ddde276 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.944427] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.944738] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.944952] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Deleting the datastore file [datastore1] 3e938bb3-54a6-49f8-809d-d5aee0349eae {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.945276] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3eac60e2-979a-4713-97c7-cb47f749f58a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.955997] env[70013]: DEBUG oslo_vmware.api [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for the task: (returnval){ [ 816.955997] env[70013]: value = "task-4230826" [ 816.955997] env[70013]: _type = "Task" [ 816.955997] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.965202] env[70013]: DEBUG oslo_vmware.api [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.978482] env[70013]: ERROR nova.scheduler.client.report [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [req-03533e7a-e6c3-41f4-b99c-78914b1502a3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-03533e7a-e6c3-41f4-b99c-78914b1502a3"}]} [ 816.978921] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.979671] env[70013]: ERROR nova.compute.manager [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 816.979671] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Traceback (most recent call last): [ 816.979671] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 816.979671] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] yield [ 816.979671] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 816.979671] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] self.set_inventory_for_provider( [ 816.979671] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 816.979671] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-03533e7a-e6c3-41f4-b99c-78914b1502a3"}]} [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] During handling of the above exception, another exception occurred: [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Traceback (most recent call last): [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] with self.rt.instance_claim(context, instance, node, allocs, [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 816.980116] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] return f(*args, **kwargs) [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] self._update(elevated, cn) [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] self._update_to_placement(context, compute_node, startup) [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] return attempt.get(self._wrap_exception) [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] six.reraise(self.value[0], self.value[1], self.value[2]) [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] raise value [ 816.980473] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] self.reportclient.update_from_provider_tree( [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] with catch_all(pd.uuid): [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] self.gen.throw(typ, value, traceback) [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] raise exception.ResourceProviderSyncFailed() [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 816.980893] env[70013]: ERROR nova.compute.manager [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] [ 816.981335] env[70013]: DEBUG nova.compute.utils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 816.981692] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.324s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.981901] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.982108] env[70013]: INFO nova.compute.manager [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Successfully reverted task state from None on failure for instance. [ 816.984500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.384s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.986100] env[70013]: INFO nova.compute.claims [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server [None req-8c799835-18d3-4b31-8dc1-ac83f3c12f5d tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server yield [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c8a0f138-d790-4bb5-b577-753219c6e400"}]} [ 816.989862] env[70013]: ERROR oslo_messaging.rpc.server [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 816.990303] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 816.990920] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 816.991458] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 816.991991] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 816.992467] env[70013]: ERROR oslo_messaging.rpc.server [ 816.992467] env[70013]: DEBUG nova.compute.manager [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Build of instance 0b42f938-50aa-4576-9231-d8f3461ee94e was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 816.992467] env[70013]: DEBUG nova.compute.manager [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 816.992772] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Acquiring lock "refresh_cache-0b42f938-50aa-4576-9231-d8f3461ee94e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.992772] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Acquired lock "refresh_cache-0b42f938-50aa-4576-9231-d8f3461ee94e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.992772] env[70013]: DEBUG nova.network.neutron [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.072400] env[70013]: INFO nova.compute.manager [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Took 36.88 seconds to build instance. [ 817.152297] env[70013]: INFO nova.compute.manager [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: f1871673-e764-460b-adce-1742e7c105c6] Took 1.05 seconds to deallocate network for instance. [ 817.171564] env[70013]: DEBUG nova.network.neutron [-] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.358568] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230823, 'name': CreateSnapshot_Task, 'duration_secs': 0.825586} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.358912] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 817.359729] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907f1406-7233-4550-939d-e01c06f3a530 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.467045] env[70013]: DEBUG oslo_vmware.api [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Task: {'id': task-4230826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169949} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.467340] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.467553] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.467794] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 817.468034] env[70013]: INFO nova.compute.manager [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Took 1.14 seconds to destroy the instance on the hypervisor. [ 817.468355] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 817.468625] env[70013]: DEBUG nova.compute.manager [-] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 817.468778] env[70013]: DEBUG nova.network.neutron [-] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.514150] env[70013]: DEBUG nova.network.neutron [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.574263] env[70013]: DEBUG oslo_concurrency.lockutils [None req-928ca46b-9e6b-4510-afa9-269afeee6990 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.138s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.642401] env[70013]: DEBUG nova.network.neutron [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.675090] env[70013]: INFO nova.compute.manager [-] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Took 1.03 seconds to deallocate network for instance. [ 817.739616] env[70013]: DEBUG nova.compute.manager [req-5a561ecc-5f0f-41d2-a117-996865a8cc8c req-0a5dc643-e096-4dd2-b57a-a3954a158fc1 service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Received event network-vif-deleted-5708c0e5-6147-4b18-b294-885738a29578 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 817.739867] env[70013]: INFO nova.compute.manager [req-5a561ecc-5f0f-41d2-a117-996865a8cc8c req-0a5dc643-e096-4dd2-b57a-a3954a158fc1 service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Neutron deleted interface 5708c0e5-6147-4b18-b294-885738a29578; detaching it from the instance and deleting it from the info cache [ 817.740026] env[70013]: DEBUG nova.network.neutron [req-5a561ecc-5f0f-41d2-a117-996865a8cc8c req-0a5dc643-e096-4dd2-b57a-a3954a158fc1 service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.887150] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 817.888116] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f563e1a1-a278-455a-93e1-baf8ecd77aa7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.906181] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 817.906181] env[70013]: value = "task-4230827" [ 817.906181] env[70013]: _type = "Task" [ 817.906181] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.916342] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230827, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.962243] env[70013]: DEBUG nova.compute.manager [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Received event network-changed-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 817.962448] env[70013]: DEBUG nova.compute.manager [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Refreshing instance network info cache due to event network-changed-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 817.962651] env[70013]: DEBUG oslo_concurrency.lockutils [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] Acquiring lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.962796] env[70013]: DEBUG oslo_concurrency.lockutils [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] Acquired lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 817.962980] env[70013]: DEBUG nova.network.neutron [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Refreshing network info cache for port 2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.015518] env[70013]: DEBUG nova.scheduler.client.report [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 818.032129] env[70013]: DEBUG nova.scheduler.client.report [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 818.032479] env[70013]: DEBUG nova.compute.provider_tree [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 818.050719] env[70013]: DEBUG nova.scheduler.client.report [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 818.076474] env[70013]: DEBUG nova.scheduler.client.report [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 818.079106] env[70013]: DEBUG nova.compute.manager [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 818.146658] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Releasing lock "refresh_cache-0b42f938-50aa-4576-9231-d8f3461ee94e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.146763] env[70013]: DEBUG nova.compute.manager [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 818.146968] env[70013]: DEBUG nova.compute.manager [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 818.147153] env[70013]: DEBUG nova.network.neutron [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.172543] env[70013]: DEBUG nova.network.neutron [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.191742] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 818.198697] env[70013]: INFO nova.scheduler.client.report [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Deleted allocations for instance f1871673-e764-460b-adce-1742e7c105c6 [ 818.218353] env[70013]: DEBUG nova.network.neutron [-] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.243038] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1fa62f3c-7856-43f8-a2d9-252b96d3551e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.259388] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698e3e53-ec02-45b1-b46c-d33d4e1ce9f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.302505] env[70013]: DEBUG nova.compute.manager [req-5a561ecc-5f0f-41d2-a117-996865a8cc8c req-0a5dc643-e096-4dd2-b57a-a3954a158fc1 service nova] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Detach interface failed, port_id=5708c0e5-6147-4b18-b294-885738a29578, reason: Instance 3e938bb3-54a6-49f8-809d-d5aee0349eae could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 818.419832] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230827, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.599493] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 818.654155] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c92a722-f8ce-4d4a-a15b-9b2ed10006c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.664029] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c407ecc0-37eb-42eb-b1e9-abedb927d74a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.695485] env[70013]: DEBUG nova.network.neutron [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.699813] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53bf382c-5774-4a25-a590-e16306991f2e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.708552] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cb16b3-1e9c-4e2a-89e2-bcb81e476a8f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.713210] env[70013]: DEBUG oslo_concurrency.lockutils [None req-66f4f9f1-5697-417d-871a-5f480fe4c29e tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "f1871673-e764-460b-adce-1742e7c105c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.989s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 818.726331] env[70013]: INFO nova.compute.manager [-] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Took 1.26 seconds to deallocate network for instance. [ 818.727256] env[70013]: DEBUG nova.compute.provider_tree [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.736720] env[70013]: DEBUG nova.network.neutron [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updated VIF entry in instance network info cache for port 2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 818.737100] env[70013]: DEBUG nova.network.neutron [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updating instance_info_cache with network_info: [{"id": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "address": "fa:16:3e:9b:0a:0d", "network": {"id": "b250add1-805e-4928-b1de-496c88097a4e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-15368811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1fa3c6a4122484c831621d0f108a5f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f01bbee7-8b9a-46be-891e-59d8142fb359", "external-id": "nsx-vlan-transportzone-145", "segmentation_id": 145, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd50eea-d7", "ovs_interfaceid": "2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.918109] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230827, 'name': CloneVM_Task} progress is 95%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.201190] env[70013]: INFO nova.compute.manager [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] [instance: 0b42f938-50aa-4576-9231-d8f3461ee94e] Took 1.05 seconds to deallocate network for instance. [ 819.217301] env[70013]: DEBUG nova.compute.manager [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 819.232613] env[70013]: DEBUG nova.scheduler.client.report [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 819.238357] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 819.239113] env[70013]: DEBUG oslo_concurrency.lockutils [req-c6f53144-22c3-433f-bb08-0cc28e91f939 req-40fba056-9daf-4b35-bffb-29be9cb18bec service nova] Releasing lock "refresh_cache-5c5fbdae-4916-4beb-ada0-57f36fb0f84c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 819.420377] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230827, 'name': CloneVM_Task, 'duration_secs': 1.283783} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.420673] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Created linked-clone VM from snapshot [ 819.422273] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbbc0e0-19c4-4c39-a711-20c35e83d4c8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.430583] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Uploading image 22b09a9a-acd0-43ee-8b3e-e07135e60269 {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 819.454662] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 819.454662] env[70013]: value = "vm-837101" [ 819.454662] env[70013]: _type = "VirtualMachine" [ 819.454662] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 819.454963] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-06aec9ed-879a-4e55-9853-7f486783b572 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.463591] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease: (returnval){ [ 819.463591] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526720f5-a050-ce25-1100-72329b193443" [ 819.463591] env[70013]: _type = "HttpNfcLease" [ 819.463591] env[70013]: } obtained for exporting VM: (result){ [ 819.463591] env[70013]: value = "vm-837101" [ 819.463591] env[70013]: _type = "VirtualMachine" [ 819.463591] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 819.463942] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the lease: (returnval){ [ 819.463942] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526720f5-a050-ce25-1100-72329b193443" [ 819.463942] env[70013]: _type = "HttpNfcLease" [ 819.463942] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 819.472684] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 819.472684] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526720f5-a050-ce25-1100-72329b193443" [ 819.472684] env[70013]: _type = "HttpNfcLease" [ 819.472684] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 819.740160] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 819.741054] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.757s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 819.741550] env[70013]: DEBUG nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 819.744713] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.566s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 819.746085] env[70013]: INFO nova.compute.claims [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.973764] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 819.973764] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526720f5-a050-ce25-1100-72329b193443" [ 819.973764] env[70013]: _type = "HttpNfcLease" [ 819.973764] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 819.974208] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 819.974208] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526720f5-a050-ce25-1100-72329b193443" [ 819.974208] env[70013]: _type = "HttpNfcLease" [ 819.974208] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 819.975049] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fb65b3-53b1-4fe2-b4f6-44498e4d8047 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.983211] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c81172-3175-172c-3962-d5bbd2889073/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 819.983397] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c81172-3175-172c-3962-d5bbd2889073/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 820.077902] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-68894e79-32e2-491b-a3a8-3b16164cc975 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.229266] env[70013]: INFO nova.scheduler.client.report [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Deleted allocations for instance 0b42f938-50aa-4576-9231-d8f3461ee94e [ 820.253734] env[70013]: DEBUG nova.compute.utils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 820.261850] env[70013]: DEBUG nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 820.261850] env[70013]: DEBUG nova.network.neutron [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.350839] env[70013]: DEBUG nova.policy [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db5dbcb88dc0460f91dfb591a556535d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6892f5b2709946579c1b58647b41ee92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 820.741404] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ea950880-8c8c-4d42-9442-d6cdf7c94f0c tempest-ServerActionsTestOtherB-500818960 tempest-ServerActionsTestOtherB-500818960-project-member] Lock "0b42f938-50aa-4576-9231-d8f3461ee94e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.361s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.758945] env[70013]: DEBUG nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 820.880889] env[70013]: DEBUG nova.network.neutron [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Successfully created port: 629bce59-9f07-4ab6-b37e-5b1fd45aa634 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.248066] env[70013]: DEBUG nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 821.395185] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff142f8-c550-4dab-8f6a-73f8861c6093 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.404891] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4573f7-3fab-4bad-946b-55cbb28cc744 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.451573] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cf4fde-1041-4938-9875-44363c4f82a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.460721] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85c0d81-7f59-4e09-a03c-acaa6b4d973f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.477817] env[70013]: DEBUG nova.compute.provider_tree [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.769721] env[70013]: DEBUG nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 821.775404] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.797201] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 821.797461] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.797621] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 821.797805] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.797951] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 821.798108] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 821.798324] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 821.798483] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 821.798648] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 821.798811] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 821.798981] env[70013]: DEBUG nova.virt.hardware [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 821.799919] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ae30c3-b424-42c7-ac31-3612545c4cf3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.808380] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050e6d1f-4dcd-42ab-b6de-83c001959061 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.982894] env[70013]: DEBUG nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 822.489950] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.744s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.489950] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 822.493428] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.201s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.493721] env[70013]: DEBUG nova.objects.instance [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lazy-loading 'resources' on Instance uuid cfe1478c-53f2-4df8-99a8-ab23209f5306 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.623171] env[70013]: DEBUG nova.compute.manager [req-a9f945d2-910a-4c0f-8633-86d02006feeb req-e910e2a5-56b0-4570-b77c-75f7f651b8a9 service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Received event network-vif-plugged-629bce59-9f07-4ab6-b37e-5b1fd45aa634 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 822.623396] env[70013]: DEBUG oslo_concurrency.lockutils [req-a9f945d2-910a-4c0f-8633-86d02006feeb req-e910e2a5-56b0-4570-b77c-75f7f651b8a9 service nova] Acquiring lock "352790fe-057f-45ee-aff0-549b456fd181-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.623783] env[70013]: DEBUG oslo_concurrency.lockutils [req-a9f945d2-910a-4c0f-8633-86d02006feeb req-e910e2a5-56b0-4570-b77c-75f7f651b8a9 service nova] Lock "352790fe-057f-45ee-aff0-549b456fd181-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.624106] env[70013]: DEBUG oslo_concurrency.lockutils [req-a9f945d2-910a-4c0f-8633-86d02006feeb req-e910e2a5-56b0-4570-b77c-75f7f651b8a9 service nova] Lock "352790fe-057f-45ee-aff0-549b456fd181-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.624408] env[70013]: DEBUG nova.compute.manager [req-a9f945d2-910a-4c0f-8633-86d02006feeb req-e910e2a5-56b0-4570-b77c-75f7f651b8a9 service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] No waiting events found dispatching network-vif-plugged-629bce59-9f07-4ab6-b37e-5b1fd45aa634 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 822.624790] env[70013]: WARNING nova.compute.manager [req-a9f945d2-910a-4c0f-8633-86d02006feeb req-e910e2a5-56b0-4570-b77c-75f7f651b8a9 service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Received unexpected event network-vif-plugged-629bce59-9f07-4ab6-b37e-5b1fd45aa634 for instance with vm_state building and task_state spawning. [ 822.774920] env[70013]: DEBUG nova.network.neutron [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Successfully updated port: 629bce59-9f07-4ab6-b37e-5b1fd45aa634 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.998928] env[70013]: DEBUG nova.compute.utils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 823.004054] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 823.004294] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.044722] env[70013]: DEBUG nova.policy [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '679f30689ad34e468723d9bdbd060e7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c143f09a5b7241fea79e9a9b8399f4e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 823.280471] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "refresh_cache-352790fe-057f-45ee-aff0-549b456fd181" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.280828] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquired lock "refresh_cache-352790fe-057f-45ee-aff0-549b456fd181" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 823.280896] env[70013]: DEBUG nova.network.neutron [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.427761] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Successfully created port: 6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.505275] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 823.536241] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24aa0f85-20d7-4bc2-b351-c0e02a803576 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.545402] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfdb1512-2797-4504-8100-af8e5b7db6bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.581723] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b64bb871-2abc-4028-bea6-9ac3dd3fb5d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.592163] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84c00f4-0f46-433f-a4f0-fdcf3f73fa0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.608681] env[70013]: DEBUG nova.compute.provider_tree [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.819867] env[70013]: DEBUG nova.network.neutron [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.974240] env[70013]: DEBUG nova.network.neutron [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Updating instance_info_cache with network_info: [{"id": "629bce59-9f07-4ab6-b37e-5b1fd45aa634", "address": "fa:16:3e:20:34:07", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap629bce59-9f", "ovs_interfaceid": "629bce59-9f07-4ab6-b37e-5b1fd45aa634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.112407] env[70013]: DEBUG nova.scheduler.client.report [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 824.477781] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Releasing lock "refresh_cache-352790fe-057f-45ee-aff0-549b456fd181" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.477781] env[70013]: DEBUG nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Instance network_info: |[{"id": "629bce59-9f07-4ab6-b37e-5b1fd45aa634", "address": "fa:16:3e:20:34:07", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap629bce59-9f", "ovs_interfaceid": "629bce59-9f07-4ab6-b37e-5b1fd45aa634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 824.478212] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:34:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '629bce59-9f07-4ab6-b37e-5b1fd45aa634', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.486642] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Creating folder: Project (6892f5b2709946579c1b58647b41ee92). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.486642] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e64c209b-0df5-419b-a165-ad6a2f6d9688 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.499645] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Created folder: Project (6892f5b2709946579c1b58647b41ee92) in parent group-v836999. [ 824.500105] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Creating folder: Instances. Parent ref: group-v837102. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.500177] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52ad54d9-c77a-4e1a-ae32-5a201e916339 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.512547] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Created folder: Instances in parent group-v837102. [ 824.513653] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 824.513978] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.514272] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1742cf64-91dd-4519-a9a0-7d5491fc6898 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.531307] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 824.540784] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.540784] env[70013]: value = "task-4230831" [ 824.540784] env[70013]: _type = "Task" [ 824.540784] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.550037] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230831, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.559571] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 824.559874] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.559987] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 824.560188] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.560338] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 824.560486] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 824.560697] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 824.560856] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 824.561074] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 824.561226] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 824.561455] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 824.562401] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc81b576-6309-4633-9ee3-cb0f8890c013 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.570825] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4ca453-2da2-44e1-8b9c-f66af4c83490 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.617220] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.124s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 824.620545] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.969s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.622430] env[70013]: INFO nova.compute.claims [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.643768] env[70013]: INFO nova.scheduler.client.report [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Deleted allocations for instance cfe1478c-53f2-4df8-99a8-ab23209f5306 [ 824.652159] env[70013]: DEBUG nova.compute.manager [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Received event network-changed-629bce59-9f07-4ab6-b37e-5b1fd45aa634 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 824.652413] env[70013]: DEBUG nova.compute.manager [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Refreshing instance network info cache due to event network-changed-629bce59-9f07-4ab6-b37e-5b1fd45aa634. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 824.652785] env[70013]: DEBUG oslo_concurrency.lockutils [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] Acquiring lock "refresh_cache-352790fe-057f-45ee-aff0-549b456fd181" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.653018] env[70013]: DEBUG oslo_concurrency.lockutils [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] Acquired lock "refresh_cache-352790fe-057f-45ee-aff0-549b456fd181" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.653205] env[70013]: DEBUG nova.network.neutron [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Refreshing network info cache for port 629bce59-9f07-4ab6-b37e-5b1fd45aa634 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.051471] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230831, 'name': CreateVM_Task, 'duration_secs': 0.467081} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.051652] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.052389] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.052546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.053190] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 825.053190] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-894e3754-89e3-40f3-938b-5c7663070984 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.064026] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 825.064026] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526c98b7-37ce-e257-3915-c854226429ac" [ 825.064026] env[70013]: _type = "Task" [ 825.064026] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.073272] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526c98b7-37ce-e257-3915-c854226429ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.165012] env[70013]: DEBUG oslo_concurrency.lockutils [None req-93ddc63c-5b66-4ce8-95c8-1c8bacd42411 tempest-TenantUsagesTestJSON-708809549 tempest-TenantUsagesTestJSON-708809549-project-member] Lock "cfe1478c-53f2-4df8-99a8-ab23209f5306" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.409s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.178947] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Successfully updated port: 6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.388687] env[70013]: DEBUG nova.network.neutron [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Updated VIF entry in instance network info cache for port 629bce59-9f07-4ab6-b37e-5b1fd45aa634. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 825.389090] env[70013]: DEBUG nova.network.neutron [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Updating instance_info_cache with network_info: [{"id": "629bce59-9f07-4ab6-b37e-5b1fd45aa634", "address": "fa:16:3e:20:34:07", "network": {"id": "f06e7b61-beea-4cb0-bfff-e11cc8c15494", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1e6c9228801846fd94db2231de2054c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap629bce59-9f", "ovs_interfaceid": "629bce59-9f07-4ab6-b37e-5b1fd45aa634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.574746] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526c98b7-37ce-e257-3915-c854226429ac, 'name': SearchDatastore_Task, 'duration_secs': 0.026834} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.575117] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.575373] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.575670] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.575784] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.575978] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.576292] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffeac7ce-d7fe-4266-a541-453709dae4ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.588050] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.588050] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.588392] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab0bdf57-b01b-4b31-9d0a-c04887b46df6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.594338] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 825.594338] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52212da3-478f-2a25-1472-25df819646c3" [ 825.594338] env[70013]: _type = "Task" [ 825.594338] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.603385] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52212da3-478f-2a25-1472-25df819646c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.682288] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "refresh_cache-02ef7875-7ec9-4409-aaa5-71ed669f3780" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.682489] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired lock "refresh_cache-02ef7875-7ec9-4409-aaa5-71ed669f3780" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.682716] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.892214] env[70013]: DEBUG oslo_concurrency.lockutils [req-2f442806-edd3-4c12-a880-d1fd2b328838 req-0f36a172-dae7-4a45-8c26-7190c335be3f service nova] Releasing lock "refresh_cache-352790fe-057f-45ee-aff0-549b456fd181" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 826.109580] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52212da3-478f-2a25-1472-25df819646c3, 'name': SearchDatastore_Task, 'duration_secs': 0.042158} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.110472] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e7f5f32-a4cf-4578-b6f9-4ce25cd9171d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.120499] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 826.120499] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527e60c3-4167-1da0-c6a0-9d3204d1f923" [ 826.120499] env[70013]: _type = "Task" [ 826.120499] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.126799] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a143789-acb9-40f2-b61a-232332d8ffc5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.132869] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527e60c3-4167-1da0-c6a0-9d3204d1f923, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.138665] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016e1239-a0af-443d-ae8e-d968e947e320 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.174750] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a5bce6-2ad1-4fd9-9fbb-989f85f4b341 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.183831] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa347aa9-e84d-4fca-a703-4bdea38f3c19 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.201340] env[70013]: DEBUG nova.compute.provider_tree [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.221400] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.364047] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Updating instance_info_cache with network_info: [{"id": "6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230", "address": "fa:16:3e:1b:6f:b8", "network": {"id": "afabfec6-94de-4cfc-8008-9c0eb77e348c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-875497650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c143f09a5b7241fea79e9a9b8399f4e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8e0b6-a8", "ovs_interfaceid": "6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.635085] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527e60c3-4167-1da0-c6a0-9d3204d1f923, 'name': SearchDatastore_Task, 'duration_secs': 0.013097} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.635435] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 826.635798] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 352790fe-057f-45ee-aff0-549b456fd181/352790fe-057f-45ee-aff0-549b456fd181.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.636126] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a67034a6-1068-4144-b971-957431e7eea8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.645681] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 826.645681] env[70013]: value = "task-4230832" [ 826.645681] env[70013]: _type = "Task" [ 826.645681] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.655843] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230832, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.684352] env[70013]: DEBUG nova.compute.manager [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Received event network-vif-plugged-6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 826.684535] env[70013]: DEBUG oslo_concurrency.lockutils [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] Acquiring lock "02ef7875-7ec9-4409-aaa5-71ed669f3780-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.684777] env[70013]: DEBUG oslo_concurrency.lockutils [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.685019] env[70013]: DEBUG oslo_concurrency.lockutils [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.685204] env[70013]: DEBUG nova.compute.manager [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] No waiting events found dispatching network-vif-plugged-6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 826.685434] env[70013]: WARNING nova.compute.manager [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Received unexpected event network-vif-plugged-6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 for instance with vm_state building and task_state spawning. [ 826.685638] env[70013]: DEBUG nova.compute.manager [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Received event network-changed-6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 826.685865] env[70013]: DEBUG nova.compute.manager [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Refreshing instance network info cache due to event network-changed-6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 826.686203] env[70013]: DEBUG oslo_concurrency.lockutils [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] Acquiring lock "refresh_cache-02ef7875-7ec9-4409-aaa5-71ed669f3780" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.705185] env[70013]: DEBUG nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 826.874551] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Releasing lock "refresh_cache-02ef7875-7ec9-4409-aaa5-71ed669f3780" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 826.874551] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Instance network_info: |[{"id": "6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230", "address": "fa:16:3e:1b:6f:b8", "network": {"id": "afabfec6-94de-4cfc-8008-9c0eb77e348c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-875497650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c143f09a5b7241fea79e9a9b8399f4e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8e0b6-a8", "ovs_interfaceid": "6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 826.874924] env[70013]: DEBUG oslo_concurrency.lockutils [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] Acquired lock "refresh_cache-02ef7875-7ec9-4409-aaa5-71ed669f3780" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 826.874924] env[70013]: DEBUG nova.network.neutron [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Refreshing network info cache for port 6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 826.880130] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:6f:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2eaecd-9701-4504-9fcb-fb1a420ead72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 826.893053] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Creating folder: Project (c143f09a5b7241fea79e9a9b8399f4e4). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 826.894753] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d10d3507-c58d-47f5-b8e6-6e9e235b8df7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.909807] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Created folder: Project (c143f09a5b7241fea79e9a9b8399f4e4) in parent group-v836999. [ 826.910053] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Creating folder: Instances. Parent ref: group-v837105. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 826.910327] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f2f6b28-b23a-44d6-94b0-6f7ec8a49e0f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.923527] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Created folder: Instances in parent group-v837105. [ 826.923825] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 826.924078] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 826.924342] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74f5c508-4f07-4300-b5e8-d95d3432aa46 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.952413] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 826.952413] env[70013]: value = "task-4230835" [ 826.952413] env[70013]: _type = "Task" [ 826.952413] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.963054] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230835, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.162501] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230832, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.211151] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.211802] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 827.215351] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.187s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.215351] env[70013]: DEBUG nova.objects.instance [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lazy-loading 'resources' on Instance uuid f12b8575-5082-4be9-9bf5-f4279860d19d {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.453208] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c81172-3175-172c-3962-d5bbd2889073/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 827.454226] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb87e25-e1e9-4d44-a79b-ba4537611095 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.472218] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c81172-3175-172c-3962-d5bbd2889073/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 827.472826] env[70013]: ERROR oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c81172-3175-172c-3962-d5bbd2889073/disk-0.vmdk due to incomplete transfer. [ 827.472952] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230835, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.473138] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b95426c4-7f4f-4730-a8c4-2ce52f9cf7d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.491496] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c81172-3175-172c-3962-d5bbd2889073/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 827.491496] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Uploaded image 22b09a9a-acd0-43ee-8b3e-e07135e60269 to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 827.494376] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 827.494564] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a46f9bb7-9f00-4b9f-9752-e4c3e845fd0c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.502652] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 827.502652] env[70013]: value = "task-4230836" [ 827.502652] env[70013]: _type = "Task" [ 827.502652] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.512606] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230836, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.664096] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230832, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.610398} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.664366] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 352790fe-057f-45ee-aff0-549b456fd181/352790fe-057f-45ee-aff0-549b456fd181.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.664555] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.664889] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1758f2d2-a094-4095-900a-ba4f036959a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.675035] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 827.675035] env[70013]: value = "task-4230837" [ 827.675035] env[70013]: _type = "Task" [ 827.675035] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.684610] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230837, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.725682] env[70013]: DEBUG nova.compute.utils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 827.728172] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 827.728504] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.774955] env[70013]: DEBUG nova.policy [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '679f30689ad34e468723d9bdbd060e7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c143f09a5b7241fea79e9a9b8399f4e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 827.797331] env[70013]: DEBUG nova.network.neutron [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Updated VIF entry in instance network info cache for port 6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 827.798124] env[70013]: DEBUG nova.network.neutron [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Updating instance_info_cache with network_info: [{"id": "6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230", "address": "fa:16:3e:1b:6f:b8", "network": {"id": "afabfec6-94de-4cfc-8008-9c0eb77e348c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-875497650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c143f09a5b7241fea79e9a9b8399f4e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8e0b6-a8", "ovs_interfaceid": "6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.970749] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230835, 'name': CreateVM_Task, 'duration_secs': 0.635133} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.971581] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 827.972101] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.972334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 827.972739] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 827.975755] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1558fa9b-0e28-411c-825e-c6217fc83d98 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.982587] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 827.982587] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521b3a9b-62f9-ca07-c32a-72077a13a0e4" [ 827.982587] env[70013]: _type = "Task" [ 827.982587] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.002022] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521b3a9b-62f9-ca07-c32a-72077a13a0e4, 'name': SearchDatastore_Task, 'duration_secs': 0.01026} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.002022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.002022] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.002022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.002290] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.002290] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.002290] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d290a529-7b01-42be-8c07-bb4b19b262e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.016124] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230836, 'name': Destroy_Task, 'duration_secs': 0.391365} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.017753] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Destroyed the VM [ 828.018045] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 828.021027] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.021027] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.021027] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4cb1baba-42d6-46e3-b70a-79d9ffa928dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.021456] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0c5ba14-2a69-4979-9e36-17ddd1821045 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.030999] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 828.030999] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5229e675-7de8-582f-fe2d-32d6a0c42e17" [ 828.030999] env[70013]: _type = "Task" [ 828.030999] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.037090] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 828.037090] env[70013]: value = "task-4230838" [ 828.037090] env[70013]: _type = "Task" [ 828.037090] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.044842] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5229e675-7de8-582f-fe2d-32d6a0c42e17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.056174] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230838, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.182810] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230837, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073305} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.186630] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.187088] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd2814d-e2eb-4b5c-b676-759d377643c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.217943] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 352790fe-057f-45ee-aff0-549b456fd181/352790fe-057f-45ee-aff0-549b456fd181.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.221221] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0853ec8b-82dc-41e7-b84d-1c957e9655f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.237669] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 828.250985] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 828.250985] env[70013]: value = "task-4230839" [ 828.250985] env[70013]: _type = "Task" [ 828.250985] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.264735] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230839, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.302324] env[70013]: DEBUG oslo_concurrency.lockutils [req-653846e5-8ce9-4364-874a-35f0e2cbd7b4 req-eda9e50a-d399-4a09-b259-513f37a57b4d service nova] Releasing lock "refresh_cache-02ef7875-7ec9-4409-aaa5-71ed669f3780" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.328516] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f6921d-b93d-415f-94c2-42b53f9c3a01 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.338604] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889fcf4e-b892-4a23-8c11-40aae3971677 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.376100] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0cf2fb-9080-448e-80ae-fa6fed53b4b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.385574] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe0ee12-1f81-4186-b133-a6166ff7e3cf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.409232] env[70013]: DEBUG nova.compute.provider_tree [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 828.438590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "77c12460-9cfa-41c9-a210-238a470d9ccd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.438590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.438590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "77c12460-9cfa-41c9-a210-238a470d9ccd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.438590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.438944] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.440555] env[70013]: INFO nova.compute.manager [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Terminating instance [ 828.541336] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5229e675-7de8-582f-fe2d-32d6a0c42e17, 'name': SearchDatastore_Task, 'duration_secs': 0.014674} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.545481] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eac94176-05c9-4cdd-a883-ae4538f17aca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.553843] env[70013]: DEBUG oslo_vmware.api [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230838, 'name': RemoveSnapshot_Task, 'duration_secs': 0.456118} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.558036] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 828.558036] env[70013]: INFO nova.compute.manager [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Took 12.76 seconds to snapshot the instance on the hypervisor. [ 828.558737] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 828.558737] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]524fde0d-8491-78cf-3a83-e1ae3efea8c4" [ 828.558737] env[70013]: _type = "Task" [ 828.558737] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.569263] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524fde0d-8491-78cf-3a83-e1ae3efea8c4, 'name': SearchDatastore_Task, 'duration_secs': 0.011175} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.569566] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.569866] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 02ef7875-7ec9-4409-aaa5-71ed669f3780/02ef7875-7ec9-4409-aaa5-71ed669f3780.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 828.570151] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1080d2f4-4019-404a-8024-f610feeb6de3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.579852] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 828.579852] env[70013]: value = "task-4230840" [ 828.579852] env[70013]: _type = "Task" [ 828.579852] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.579852] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Successfully created port: addd72e9-8e61-463a-ab33-7872333bd242 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.592786] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.766714] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.948024] env[70013]: DEBUG nova.compute.manager [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 828.948024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 828.948024] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5fba77-769e-4ba2-bc8b-b68de0d259e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.952581] env[70013]: ERROR nova.scheduler.client.report [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [req-5b10b7b6-fd96-4bb9-b6c1-f8a8ba62e9ef] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5b10b7b6-fd96-4bb9-b6c1-f8a8ba62e9ef"}]} [ 828.953372] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.954334] env[70013]: ERROR nova.compute.manager [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 828.954334] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Traceback (most recent call last): [ 828.954334] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 828.954334] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] yield [ 828.954334] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 828.954334] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self.set_inventory_for_provider( [ 828.954334] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 828.954334] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5b10b7b6-fd96-4bb9-b6c1-f8a8ba62e9ef"}]} [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] During handling of the above exception, another exception occurred: [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Traceback (most recent call last): [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self._delete_instance(context, instance, bdms) [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 828.954671] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self._complete_deletion(context, instance) [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self._update_resource_tracker(context, instance) [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self.rt.update_usage(context, instance, instance.node) [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] return f(*args, **kwargs) [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self._update(context.elevated(), self.compute_nodes[nodename]) [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self._update_to_placement(context, compute_node, startup) [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 828.955033] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] return attempt.get(self._wrap_exception) [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] raise value [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self.reportclient.update_from_provider_tree( [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] with catch_all(pd.uuid): [ 828.955387] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 828.955870] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] self.gen.throw(typ, value, traceback) [ 828.955870] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 828.955870] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] raise exception.ResourceProviderSyncFailed() [ 828.955870] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 828.955870] env[70013]: ERROR nova.compute.manager [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] [ 828.961403] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.329s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.961403] env[70013]: INFO nova.compute.claims [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.973620] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 828.974193] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8bd330b-2c37-4f95-a088-7dc1671b2662 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.057027] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.057027] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.057027] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleting the datastore file [datastore2] 77c12460-9cfa-41c9-a210-238a470d9ccd {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.057027] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-172e00fb-4c26-4087-9432-fdfd6fcc1fce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.060955] env[70013]: DEBUG nova.compute.manager [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Instance disappeared during snapshot {{(pid=70013) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4595}} [ 829.067197] env[70013]: DEBUG oslo_vmware.api [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 829.067197] env[70013]: value = "task-4230842" [ 829.067197] env[70013]: _type = "Task" [ 829.067197] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.079534] env[70013]: DEBUG oslo_vmware.api [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.089847] env[70013]: DEBUG nova.compute.manager [None req-f5bdb4b3-a0d3-448f-a813-a2a68307e3c8 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image not found during clean up 22b09a9a-acd0-43ee-8b3e-e07135e60269 {{(pid=70013) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4601}} [ 829.096673] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230840, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.255749] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 829.274991] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230839, 'name': ReconfigVM_Task, 'duration_secs': 0.871342} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.275900] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 352790fe-057f-45ee-aff0-549b456fd181/352790fe-057f-45ee-aff0-549b456fd181.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.277800] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a833f437-8c58-4cda-acee-28e00d190119 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.285752] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 829.285752] env[70013]: value = "task-4230843" [ 829.285752] env[70013]: _type = "Task" [ 829.285752] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.299231] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 829.299519] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.299689] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 829.299874] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.300528] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 829.300754] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 829.301291] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 829.301291] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 829.301483] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 829.301483] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 829.301665] env[70013]: DEBUG nova.virt.hardware [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 829.301977] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230843, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.302778] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50a493a-a89f-42a3-88f3-75d474d81d51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.312319] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30134cf-ea2d-4e42-81da-1c938f3da2aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.319859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "eab0c393-4d3d-4659-8225-c82ddadb330b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.319950] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.468628] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.087s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.579764] env[70013]: DEBUG oslo_vmware.api [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181454} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.580066] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.580258] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.580446] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.580630] env[70013]: INFO nova.compute.manager [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Took 0.63 seconds to destroy the instance on the hypervisor. [ 829.580866] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 829.581063] env[70013]: DEBUG nova.compute.manager [-] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 829.581161] env[70013]: DEBUG nova.network.neutron [-] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.592500] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544722} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.592883] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 02ef7875-7ec9-4409-aaa5-71ed669f3780/02ef7875-7ec9-4409-aaa5-71ed669f3780.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 829.593205] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 829.593547] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-474279cd-379b-49cc-b4de-e9a55e64b1e1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.601752] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 829.601752] env[70013]: value = "task-4230844" [ 829.601752] env[70013]: _type = "Task" [ 829.601752] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.612809] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.799025] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230843, 'name': Rename_Task, 'duration_secs': 0.245005} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.799294] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.799383] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d8da31b6-ac41-421f-a0d6-db5e636def79 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.807197] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 829.807197] env[70013]: value = "task-4230845" [ 829.807197] env[70013]: _type = "Task" [ 829.807197] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.816200] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230845, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.997525] env[70013]: DEBUG nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 830.016313] env[70013]: DEBUG nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 830.016540] env[70013]: DEBUG nova.compute.provider_tree [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.029786] env[70013]: DEBUG nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 830.051981] env[70013]: DEBUG nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 830.119365] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072259} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.119789] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.121915] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04483655-3035-48e1-9ff6-69c2d5fe65bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.154425] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 02ef7875-7ec9-4409-aaa5-71ed669f3780/02ef7875-7ec9-4409-aaa5-71ed669f3780.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.157520] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b734f4b7-c525-477d-a084-9df66733d353 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.184135] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 830.184135] env[70013]: value = "task-4230846" [ 830.184135] env[70013]: _type = "Task" [ 830.184135] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.200185] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230846, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.207387] env[70013]: DEBUG nova.compute.manager [req-4794252a-0a2d-44b9-9681-b862300f355b req-b014c9c3-77bc-4187-a697-82a005c9adfb service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Received event network-vif-deleted-7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 830.207618] env[70013]: INFO nova.compute.manager [req-4794252a-0a2d-44b9-9681-b862300f355b req-b014c9c3-77bc-4187-a697-82a005c9adfb service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Neutron deleted interface 7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3; detaching it from the instance and deleting it from the info cache [ 830.207838] env[70013]: DEBUG nova.network.neutron [req-4794252a-0a2d-44b9-9681-b862300f355b req-b014c9c3-77bc-4187-a697-82a005c9adfb service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.320715] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230845, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.432627] env[70013]: DEBUG nova.compute.manager [req-e5530644-df0d-4f26-ba45-84268429c251 req-9f9037de-975c-409f-aa3d-c49ec33a38f2 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Received event network-vif-plugged-addd72e9-8e61-463a-ab33-7872333bd242 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 830.432884] env[70013]: DEBUG oslo_concurrency.lockutils [req-e5530644-df0d-4f26-ba45-84268429c251 req-9f9037de-975c-409f-aa3d-c49ec33a38f2 service nova] Acquiring lock "b1d844cc-dcdc-47e2-85c3-484d0862be60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.433137] env[70013]: DEBUG oslo_concurrency.lockutils [req-e5530644-df0d-4f26-ba45-84268429c251 req-9f9037de-975c-409f-aa3d-c49ec33a38f2 service nova] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.433326] env[70013]: DEBUG oslo_concurrency.lockutils [req-e5530644-df0d-4f26-ba45-84268429c251 req-9f9037de-975c-409f-aa3d-c49ec33a38f2 service nova] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.433490] env[70013]: DEBUG nova.compute.manager [req-e5530644-df0d-4f26-ba45-84268429c251 req-9f9037de-975c-409f-aa3d-c49ec33a38f2 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] No waiting events found dispatching network-vif-plugged-addd72e9-8e61-463a-ab33-7872333bd242 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 830.433645] env[70013]: WARNING nova.compute.manager [req-e5530644-df0d-4f26-ba45-84268429c251 req-9f9037de-975c-409f-aa3d-c49ec33a38f2 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Received unexpected event network-vif-plugged-addd72e9-8e61-463a-ab33-7872333bd242 for instance with vm_state building and task_state spawning. [ 830.510664] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.510960] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.621295] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Successfully updated port: addd72e9-8e61-463a-ab33-7872333bd242 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.684764] env[70013]: DEBUG nova.network.neutron [-] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.701766] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.713817] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-159e2e82-644d-47ad-b14f-dd710a791cb9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.725094] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc068e3-a353-40aa-8a28-5db234c8eec1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.738878] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9f075f-0c7d-442c-9356-42502fda9a23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.751959] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a64b540-b22f-47ec-b81b-d8ab100e6481 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.770552] env[70013]: DEBUG nova.compute.manager [req-4794252a-0a2d-44b9-9681-b862300f355b req-b014c9c3-77bc-4187-a697-82a005c9adfb service nova] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Detach interface failed, port_id=7feff77e-7bce-4e3d-8d5a-36ed30ea0fc3, reason: Instance 77c12460-9cfa-41c9-a210-238a470d9ccd could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 830.800524] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867cdea5-cf73-4063-9956-5030f08ec7ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.814419] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c954a0-3424-4244-9211-875224063ad0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.830032] env[70013]: DEBUG nova.compute.provider_tree [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.835674] env[70013]: DEBUG oslo_vmware.api [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230845, 'name': PowerOnVM_Task, 'duration_secs': 0.884886} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.835674] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.835674] env[70013]: INFO nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Took 9.07 seconds to spawn the instance on the hypervisor. [ 830.835674] env[70013]: DEBUG nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.836618] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d5aaf3-2a32-436e-a08c-a5e227134c31 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.988668] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.014048] env[70013]: DEBUG nova.compute.utils [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 831.129527] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "refresh_cache-b1d844cc-dcdc-47e2-85c3-484d0862be60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.129762] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired lock "refresh_cache-b1d844cc-dcdc-47e2-85c3-484d0862be60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.129948] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.192994] env[70013]: INFO nova.compute.manager [-] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Took 1.61 seconds to deallocate network for instance. [ 831.202600] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230846, 'name': ReconfigVM_Task, 'duration_secs': 0.786034} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.203259] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 02ef7875-7ec9-4409-aaa5-71ed669f3780/02ef7875-7ec9-4409-aaa5-71ed669f3780.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.204283] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9daee76c-5e92-4459-89f2-b15677a98ccc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.213697] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 831.213697] env[70013]: value = "task-4230847" [ 831.213697] env[70013]: _type = "Task" [ 831.213697] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.227423] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230847, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.356347] env[70013]: ERROR nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [req-aaaefc83-0cb2-49f5-b999-72822e6ff986] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-aaaefc83-0cb2-49f5-b999-72822e6ff986"}]} [ 831.357522] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.357988] env[70013]: ERROR nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 831.357988] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] Traceback (most recent call last): [ 831.357988] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 831.357988] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] yield [ 831.357988] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 831.357988] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] self.set_inventory_for_provider( [ 831.357988] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 831.357988] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-aaaefc83-0cb2-49f5-b999-72822e6ff986"}]} [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] During handling of the above exception, another exception occurred: [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] Traceback (most recent call last): [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] with self.rt.instance_claim(context, instance, node, allocs, [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 831.358567] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] return f(*args, **kwargs) [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] self._update(elevated, cn) [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] self._update_to_placement(context, compute_node, startup) [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] return attempt.get(self._wrap_exception) [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] six.reraise(self.value[0], self.value[1], self.value[2]) [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] raise value [ 831.358914] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] self.reportclient.update_from_provider_tree( [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] with catch_all(pd.uuid): [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] self.gen.throw(typ, value, traceback) [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] raise exception.ResourceProviderSyncFailed() [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 831.359336] env[70013]: ERROR nova.compute.manager [instance: c0e830f3-e28b-4637-bcce-601596e82360] [ 831.359630] env[70013]: DEBUG nova.compute.utils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 831.360188] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.875s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.369234] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Build of instance c0e830f3-e28b-4637-bcce-601596e82360 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 831.370087] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 831.370087] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "refresh_cache-c0e830f3-e28b-4637-bcce-601596e82360" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.370197] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired lock "refresh_cache-c0e830f3-e28b-4637-bcce-601596e82360" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.370297] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.380306] env[70013]: INFO nova.compute.manager [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Took 37.82 seconds to build instance. [ 831.520204] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.690561] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.707194] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.724439] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230847, 'name': Rename_Task, 'duration_secs': 0.297603} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.724679] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.724940] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7b0ab97-f380-4120-a33c-f13389252bd4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.735417] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 831.735417] env[70013]: value = "task-4230848" [ 831.735417] env[70013]: _type = "Task" [ 831.735417] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.744701] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230848, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.893693] env[70013]: DEBUG oslo_concurrency.lockutils [None req-301c45a6-55d4-4f2e-bc21-1b06ffad2aa9 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "352790fe-057f-45ee-aff0-549b456fd181" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.867s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.917395] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.948763] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Updating instance_info_cache with network_info: [{"id": "addd72e9-8e61-463a-ab33-7872333bd242", "address": "fa:16:3e:82:e4:75", "network": {"id": "afabfec6-94de-4cfc-8008-9c0eb77e348c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-875497650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c143f09a5b7241fea79e9a9b8399f4e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaddd72e9-8e", "ovs_interfaceid": "addd72e9-8e61-463a-ab33-7872333bd242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.999154] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.247939] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230848, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.395724] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Applying migration context for instance 3b29c99a-0d55-40b1-a155-199b1ced2146 as it has an incoming, in-progress migration 2b86447b-3ba9-457a-94d4-7269f0b254ab. Migration status is error {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 832.398338] env[70013]: DEBUG nova.compute.manager [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 832.452779] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Releasing lock "refresh_cache-b1d844cc-dcdc-47e2-85c3-484d0862be60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.453151] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Instance network_info: |[{"id": "addd72e9-8e61-463a-ab33-7872333bd242", "address": "fa:16:3e:82:e4:75", "network": {"id": "afabfec6-94de-4cfc-8008-9c0eb77e348c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-875497650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c143f09a5b7241fea79e9a9b8399f4e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaddd72e9-8e", "ovs_interfaceid": "addd72e9-8e61-463a-ab33-7872333bd242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 832.454076] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:e4:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2eaecd-9701-4504-9fcb-fb1a420ead72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'addd72e9-8e61-463a-ab33-7872333bd242', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.466693] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 832.467403] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 832.467491] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d07e5ab3-751d-47f3-8b04-795cb18dec53 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.495225] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.495225] env[70013]: value = "task-4230849" [ 832.495225] env[70013]: _type = "Task" [ 832.495225] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.504564] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Releasing lock "refresh_cache-c0e830f3-e28b-4637-bcce-601596e82360" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.504829] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 832.505047] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 832.505228] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.506892] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230849, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.522999] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.618143] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.619883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.619883] env[70013]: INFO nova.compute.manager [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Attaching volume 8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0 to /dev/sdb [ 832.672954] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062744ce-0c0c-4dfc-b7b9-1e5f1808716b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.682942] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34189fb2-0197-433d-b951-1cbf5faf5d87 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.702679] env[70013]: DEBUG nova.virt.block_device [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating existing volume attachment record: 37838b23-df48-4dfe-b271-30bfa3fad2a3 {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 832.748705] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230848, 'name': PowerOnVM_Task, 'duration_secs': 0.690015} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.749392] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.749392] env[70013]: INFO nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Took 8.22 seconds to spawn the instance on the hypervisor. [ 832.749546] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 832.750394] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd50762e-81f4-4547-bced-3489eae583c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.766768] env[70013]: DEBUG nova.compute.manager [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Received event network-changed-addd72e9-8e61-463a-ab33-7872333bd242 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 832.767084] env[70013]: DEBUG nova.compute.manager [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Refreshing instance network info cache due to event network-changed-addd72e9-8e61-463a-ab33-7872333bd242. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 832.767427] env[70013]: DEBUG oslo_concurrency.lockutils [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] Acquiring lock "refresh_cache-b1d844cc-dcdc-47e2-85c3-484d0862be60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.767606] env[70013]: DEBUG oslo_concurrency.lockutils [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] Acquired lock "refresh_cache-b1d844cc-dcdc-47e2-85c3-484d0862be60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 832.767744] env[70013]: DEBUG nova.network.neutron [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Refreshing network info cache for port addd72e9-8e61-463a-ab33-7872333bd242 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.943675] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.008149] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230849, 'name': CreateVM_Task, 'duration_secs': 0.406036} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.008149] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.008149] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.008149] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.008149] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 833.009194] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fa91e50-1ef8-432a-9e1e-fca8fd9e479b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.013626] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 833.013626] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529cf344-7a9b-14e0-713a-f6186ad55690" [ 833.013626] env[70013]: _type = "Task" [ 833.013626] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.026693] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.028789] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529cf344-7a9b-14e0-713a-f6186ad55690, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.276743] env[70013]: INFO nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Took 38.12 seconds to build instance. [ 833.282154] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.282154] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.320624] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "2aefaa91-c439-486a-8b19-c6f45f52583f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.320885] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.529051] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529cf344-7a9b-14e0-713a-f6186ad55690, 'name': SearchDatastore_Task, 'duration_secs': 0.021166} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.529051] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.529051] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.529051] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.529487] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.529487] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.529487] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c63307d-29de-476c-af3d-ed3932a8fabc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.531338] env[70013]: INFO nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: c0e830f3-e28b-4637-bcce-601596e82360] Took 1.03 seconds to deallocate network for instance. [ 833.545132] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.545621] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.546185] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b75f0343-1c93-4914-950d-860bc1e74d69 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.554409] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 833.554409] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]524fdbd6-b1d4-0191-1e11-8ce7593fac1d" [ 833.554409] env[70013]: _type = "Task" [ 833.554409] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.566304] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524fdbd6-b1d4-0191-1e11-8ce7593fac1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.653648] env[70013]: DEBUG nova.network.neutron [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Updated VIF entry in instance network info cache for port addd72e9-8e61-463a-ab33-7872333bd242. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 833.654063] env[70013]: DEBUG nova.network.neutron [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Updating instance_info_cache with network_info: [{"id": "addd72e9-8e61-463a-ab33-7872333bd242", "address": "fa:16:3e:82:e4:75", "network": {"id": "afabfec6-94de-4cfc-8008-9c0eb77e348c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-875497650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c143f09a5b7241fea79e9a9b8399f4e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaddd72e9-8e", "ovs_interfaceid": "addd72e9-8e61-463a-ab33-7872333bd242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.738275] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "352790fe-057f-45ee-aff0-549b456fd181" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.738580] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "352790fe-057f-45ee-aff0-549b456fd181" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.738796] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "352790fe-057f-45ee-aff0-549b456fd181-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.738980] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "352790fe-057f-45ee-aff0-549b456fd181-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.739184] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "352790fe-057f-45ee-aff0-549b456fd181-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.741445] env[70013]: INFO nova.compute.manager [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Terminating instance [ 833.767776] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-9a0ecd6c-5853-446d-8f41-b2ee51e5259a-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.767776] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-9a0ecd6c-5853-446d-8f41-b2ee51e5259a-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.767776] env[70013]: DEBUG nova.objects.instance [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'flavor' on Instance uuid 9a0ecd6c-5853-446d-8f41-b2ee51e5259a {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.783239] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.020s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 834.073056] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524fdbd6-b1d4-0191-1e11-8ce7593fac1d, 'name': SearchDatastore_Task, 'duration_secs': 0.019306} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.074323] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76fcff90-cd6a-4b89-a245-aecbd7affd53 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.083732] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 834.083732] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f30d89-384f-f17b-95f2-52d1b7070862" [ 834.083732] env[70013]: _type = "Task" [ 834.083732] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.094168] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f30d89-384f-f17b-95f2-52d1b7070862, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.158810] env[70013]: DEBUG oslo_concurrency.lockutils [req-a97a7d12-b993-4e50-96eb-64a2644565b2 req-cc92b813-0f55-4b28-a477-d573dd9ec940 service nova] Releasing lock "refresh_cache-b1d844cc-dcdc-47e2-85c3-484d0862be60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.246544] env[70013]: DEBUG nova.compute.manager [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 834.246784] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.248152] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73810f86-094e-45b9-a268-3a653fed2ef2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.256319] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.256703] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-364a9138-3087-4bde-9da5-da8b5e7e659f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.264630] env[70013]: DEBUG oslo_vmware.api [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 834.264630] env[70013]: value = "task-4230853" [ 834.264630] env[70013]: _type = "Task" [ 834.264630] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.277701] env[70013]: DEBUG oslo_vmware.api [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230853, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.286824] env[70013]: DEBUG nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 834.364141] env[70013]: DEBUG nova.objects.instance [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'pci_requests' on Instance uuid 9a0ecd6c-5853-446d-8f41-b2ee51e5259a {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.420938] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Skipping migration as instance is neither resizing nor live-migrating. {{(pid=70013) _update_usage_from_migrations /opt/stack/nova/nova/compute/resource_tracker.py:1593}} [ 834.454795] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 28a56a75-9b30-4121-8252-a9e57287441c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.454972] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 63c3b57c-022a-4eee-b215-6dd89142e659 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.455102] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.455229] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance fc473d40-b57b-437e-9511-58a0ba700a69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.455351] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.455652] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.455799] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance b23b4e67-67a6-4117-9a62-3c0dcd25e828 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.455920] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 3b29c99a-0d55-40b1-a155-199b1ced2146 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.456120] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 3e938bb3-54a6-49f8-809d-d5aee0349eae is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 834.456198] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a0ecd6c-5853-446d-8f41-b2ee51e5259a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.456316] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 09544922-6dc4-48b5-8cfd-e91e7f74c13f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.456425] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2c0e4326-c33a-42bb-b793-a100157b1c03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.456538] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 85488f5e-bd8a-4bcf-8420-744c54176c2e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.456650] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 67b90fec-a6fd-471d-a425-82bc63de572a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.456780] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2ee62ad3-0125-47dc-b163-7d15b6a17c8a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 834.456904] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 77c12460-9cfa-41c9-a210-238a470d9ccd is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 834.457037] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ac8aa41c-375a-4231-9a5d-12d6c4bba44c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 834.457158] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.457270] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 352790fe-057f-45ee-aff0-549b456fd181 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.457411] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 02ef7875-7ec9-4409-aaa5-71ed669f3780 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.457528] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance b1d844cc-dcdc-47e2-85c3-484d0862be60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.457640] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c0e830f3-e28b-4637-bcce-601596e82360 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 834.564587] env[70013]: INFO nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Deleted allocations for instance c0e830f3-e28b-4637-bcce-601596e82360 [ 834.597230] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f30d89-384f-f17b-95f2-52d1b7070862, 'name': SearchDatastore_Task, 'duration_secs': 0.046648} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.597547] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.597821] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] b1d844cc-dcdc-47e2-85c3-484d0862be60/b1d844cc-dcdc-47e2-85c3-484d0862be60.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.598099] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad9c838f-c5f8-462b-88ac-eb9632fe7077 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.604958] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 834.604958] env[70013]: value = "task-4230854" [ 834.604958] env[70013]: _type = "Task" [ 834.604958] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.614198] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.776629] env[70013]: DEBUG oslo_vmware.api [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230853, 'name': PowerOffVM_Task, 'duration_secs': 0.260936} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.776907] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 834.777550] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 834.778244] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39ad15e5-affd-413a-b5d6-fddc22072e60 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.808177] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 834.867544] env[70013]: DEBUG nova.objects.base [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Object Instance<9a0ecd6c-5853-446d-8f41-b2ee51e5259a> lazy-loaded attributes: flavor,pci_requests {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 834.868310] env[70013]: DEBUG nova.network.neutron [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 834.916041] env[70013]: DEBUG nova.policy [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 834.962082] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f715cf45-81f1-4979-abfc-455b3e64f00e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 835.076370] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "c0e830f3-e28b-4637-bcce-601596e82360" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.239s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.119425] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230854, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.306379] env[70013]: DEBUG nova.network.neutron [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Successfully created port: 022596bd-d9fb-42f4-b6ee-29eb5192e817 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.467579] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance be9a565b-c74c-442e-8087-cc4579b1c8e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 835.577014] env[70013]: DEBUG nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 835.616356] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551193} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.619088] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] b1d844cc-dcdc-47e2-85c3-484d0862be60/b1d844cc-dcdc-47e2-85c3-484d0862be60.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.619088] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.619088] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45c43f1d-0a59-4043-813e-d2ce9094c457 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.625253] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 835.625253] env[70013]: value = "task-4230857" [ 835.625253] env[70013]: _type = "Task" [ 835.625253] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.636218] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.707269] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.707512] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.707710] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Deleting the datastore file [datastore1] 352790fe-057f-45ee-aff0-549b456fd181 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.708038] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2445f389-9183-4f5c-9dd7-67a0b51858b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.714680] env[70013]: DEBUG oslo_vmware.api [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for the task: (returnval){ [ 835.714680] env[70013]: value = "task-4230858" [ 835.714680] env[70013]: _type = "Task" [ 835.714680] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.725652] env[70013]: DEBUG oslo_vmware.api [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.972043] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 836.102137] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.136308] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230857, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076848} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.136594] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.137420] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414867f5-8d4a-4977-a044-679d7fe58798 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.162775] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] b1d844cc-dcdc-47e2-85c3-484d0862be60/b1d844cc-dcdc-47e2-85c3-484d0862be60.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.163978] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-495a14da-0a6b-445f-9fd7-e3acf20d8436 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.188616] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 836.188616] env[70013]: value = "task-4230859" [ 836.188616] env[70013]: _type = "Task" [ 836.188616] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.198182] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230859, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.228041] env[70013]: DEBUG oslo_vmware.api [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.475954] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 57485f78-b97e-423c-b1f1-8648e963e5dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 836.703358] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230859, 'name': ReconfigVM_Task, 'duration_secs': 0.313222} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.703358] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Reconfigured VM instance instance-00000028 to attach disk [datastore2] b1d844cc-dcdc-47e2-85c3-484d0862be60/b1d844cc-dcdc-47e2-85c3-484d0862be60.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.703358] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f36f9ee6-f998-4a33-a9cb-f4fa608776e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.710164] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 836.710164] env[70013]: value = "task-4230860" [ 836.710164] env[70013]: _type = "Task" [ 836.710164] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.718993] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230860, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.727423] env[70013]: DEBUG oslo_vmware.api [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Task: {'id': task-4230858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.740319} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.727746] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.727911] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.728148] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.728386] env[70013]: INFO nova.compute.manager [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Took 2.48 seconds to destroy the instance on the hypervisor. [ 836.728648] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 836.728847] env[70013]: DEBUG nova.compute.manager [-] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 836.728945] env[70013]: DEBUG nova.network.neutron [-] [instance: 352790fe-057f-45ee-aff0-549b456fd181] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.887223] env[70013]: DEBUG nova.network.neutron [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Successfully updated port: 022596bd-d9fb-42f4-b6ee-29eb5192e817 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.979903] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 477bf0d8-e988-43e9-a573-994a0d08df0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 837.221711] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230860, 'name': Rename_Task, 'duration_secs': 0.168681} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.222048] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.222296] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83f6ac8e-88ab-40d0-aec7-56d84a79649e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.230073] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 837.230073] env[70013]: value = "task-4230861" [ 837.230073] env[70013]: _type = "Task" [ 837.230073] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.240654] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230861, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.260639] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 837.261733] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837110', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'name': 'volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2c0e4326-c33a-42bb-b793-a100157b1c03', 'attached_at': '', 'detached_at': '', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'serial': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 837.261877] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ae0f30-768b-4242-b382-6d0f93d3ec73 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.282398] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84881da8-2ecb-49a6-9609-c4184585edba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.307812] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Reconfiguring VM instance instance-00000017 to attach disk [localhost-esx-install-datastore (2)] volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0/volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.308162] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-159926fc-df6f-419e-b574-6506fa0d4621 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.327639] env[70013]: DEBUG oslo_vmware.api [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 837.327639] env[70013]: value = "task-4230862" [ 837.327639] env[70013]: _type = "Task" [ 837.327639] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.336416] env[70013]: DEBUG oslo_vmware.api [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230862, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.390057] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.390108] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 837.390282] env[70013]: DEBUG nova.network.neutron [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.395707] env[70013]: DEBUG nova.compute.manager [req-c1bf81e6-92c6-4d13-ad79-0feac0562e29 req-3ab52046-5c74-4a4f-aa3c-612dbbf1146a service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received event network-vif-plugged-022596bd-d9fb-42f4-b6ee-29eb5192e817 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 837.396043] env[70013]: DEBUG oslo_concurrency.lockutils [req-c1bf81e6-92c6-4d13-ad79-0feac0562e29 req-3ab52046-5c74-4a4f-aa3c-612dbbf1146a service nova] Acquiring lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.396308] env[70013]: DEBUG oslo_concurrency.lockutils [req-c1bf81e6-92c6-4d13-ad79-0feac0562e29 req-3ab52046-5c74-4a4f-aa3c-612dbbf1146a service nova] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.396485] env[70013]: DEBUG oslo_concurrency.lockutils [req-c1bf81e6-92c6-4d13-ad79-0feac0562e29 req-3ab52046-5c74-4a4f-aa3c-612dbbf1146a service nova] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.396655] env[70013]: DEBUG nova.compute.manager [req-c1bf81e6-92c6-4d13-ad79-0feac0562e29 req-3ab52046-5c74-4a4f-aa3c-612dbbf1146a service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] No waiting events found dispatching network-vif-plugged-022596bd-d9fb-42f4-b6ee-29eb5192e817 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 837.396823] env[70013]: WARNING nova.compute.manager [req-c1bf81e6-92c6-4d13-ad79-0feac0562e29 req-3ab52046-5c74-4a4f-aa3c-612dbbf1146a service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received unexpected event network-vif-plugged-022596bd-d9fb-42f4-b6ee-29eb5192e817 for instance with vm_state active and task_state None. [ 837.484491] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 0b0b2769-74ba-4ba2-9e8b-148cc2976321 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 837.554489] env[70013]: DEBUG nova.network.neutron [-] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.583314] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "b1d844cc-dcdc-47e2-85c3-484d0862be60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.595326] env[70013]: DEBUG nova.compute.manager [req-2954b85c-c0ab-4185-9303-2bafcf9c2b75 req-129c0a6e-f208-48e2-b0c1-77b1a014635c service nova] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Received event network-vif-deleted-629bce59-9f07-4ab6-b37e-5b1fd45aa634 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 837.742437] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230861, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.838088] env[70013]: DEBUG oslo_vmware.api [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230862, 'name': ReconfigVM_Task, 'duration_secs': 0.243046} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.838388] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Reconfigured VM instance instance-00000017 to attach disk [localhost-esx-install-datastore (2)] volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0/volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.843287] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-638a6756-926a-4ee0-812e-e5323c04ce23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.858979] env[70013]: DEBUG oslo_vmware.api [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 837.858979] env[70013]: value = "task-4230863" [ 837.858979] env[70013]: _type = "Task" [ 837.858979] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.870228] env[70013]: DEBUG oslo_vmware.api [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230863, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.927920] env[70013]: WARNING nova.network.neutron [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] b160f70a-0663-495b-bd35-440c047157c8 already exists in list: networks containing: ['b160f70a-0663-495b-bd35-440c047157c8']. ignoring it [ 837.987716] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 838.057779] env[70013]: INFO nova.compute.manager [-] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Took 1.33 seconds to deallocate network for instance. [ 838.163949] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "02ef7875-7ec9-4409-aaa5-71ed669f3780" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.164220] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.164427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "02ef7875-7ec9-4409-aaa5-71ed669f3780-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.164608] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.164775] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.167287] env[70013]: INFO nova.compute.manager [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Terminating instance [ 838.244096] env[70013]: DEBUG oslo_vmware.api [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230861, 'name': PowerOnVM_Task, 'duration_secs': 0.557041} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.244398] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.244597] env[70013]: INFO nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Took 8.99 seconds to spawn the instance on the hypervisor. [ 838.244785] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 838.245661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a347f1b2-4e3e-4171-9089-4fcd6efda073 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.275051] env[70013]: DEBUG nova.network.neutron [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "022596bd-d9fb-42f4-b6ee-29eb5192e817", "address": "fa:16:3e:ba:06:18", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022596bd-d9", "ovs_interfaceid": "022596bd-d9fb-42f4-b6ee-29eb5192e817", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.369861] env[70013]: DEBUG oslo_vmware.api [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4230863, 'name': ReconfigVM_Task, 'duration_secs': 0.228097} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.371603] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837110', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'name': 'volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2c0e4326-c33a-42bb-b793-a100157b1c03', 'attached_at': '', 'detached_at': '', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'serial': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 838.491049] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance a53389ce-d0c2-48ae-83e3-ee58fbdbfb01 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 838.566674] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.671789] env[70013]: DEBUG nova.compute.manager [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 838.672037] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 838.672990] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b02356-d126-49ae-84e0-867b8d51d914 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.681230] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 838.681480] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7602c8e-0a6c-442d-9ade-42ce73356f10 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.688059] env[70013]: DEBUG oslo_vmware.api [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 838.688059] env[70013]: value = "task-4230864" [ 838.688059] env[70013]: _type = "Task" [ 838.688059] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.700505] env[70013]: DEBUG oslo_vmware.api [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230864, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.756976] env[70013]: DEBUG nova.compute.utils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Conflict updating instance b1d844cc-dcdc-47e2-85c3-484d0862be60. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 838.758900] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Instance disappeared during build. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2520}} [ 838.759093] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 838.759266] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 838.759443] env[70013]: DEBUG nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 838.759733] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 838.779118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 838.779995] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.780235] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.781167] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190850cc-da55-4d51-88c0-555facb43ab8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.801984] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 838.802263] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.802439] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 838.802654] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.802827] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 838.802991] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 838.803289] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 838.803623] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 838.803695] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 838.803850] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 838.804059] env[70013]: DEBUG nova.virt.hardware [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 838.811732] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Reconfiguring VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 838.814324] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59c0e83c-cecd-410b-9d77-169de2efa8d8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.832941] env[70013]: DEBUG oslo_vmware.api [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 838.832941] env[70013]: value = "task-4230865" [ 838.832941] env[70013]: _type = "Task" [ 838.832941] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.842562] env[70013]: DEBUG oslo_vmware.api [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230865, 'name': ReconfigVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.993972] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 839.198744] env[70013]: DEBUG oslo_vmware.api [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230864, 'name': PowerOffVM_Task, 'duration_secs': 0.196709} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.199071] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.199255] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.199527] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3972c966-a6ee-4af5-9f12-c9b5de30143b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.269565] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.269944] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.270093] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Deleting the datastore file [datastore1] 02ef7875-7ec9-4409-aaa5-71ed669f3780 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.270446] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45d77a5f-b093-452f-98d2-4e935d2a8ff0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.278405] env[70013]: DEBUG oslo_vmware.api [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 839.278405] env[70013]: value = "task-4230867" [ 839.278405] env[70013]: _type = "Task" [ 839.278405] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.288665] env[70013]: DEBUG oslo_vmware.api [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.344401] env[70013]: DEBUG oslo_vmware.api [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230865, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.432037] env[70013]: DEBUG nova.objects.instance [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lazy-loading 'flavor' on Instance uuid 2c0e4326-c33a-42bb-b793-a100157b1c03 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.498568] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 839.633131] env[70013]: DEBUG nova.network.neutron [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.778515] env[70013]: DEBUG nova.compute.manager [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received event network-changed-022596bd-d9fb-42f4-b6ee-29eb5192e817 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 839.778686] env[70013]: DEBUG nova.compute.manager [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Refreshing instance network info cache due to event network-changed-022596bd-d9fb-42f4-b6ee-29eb5192e817. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 839.778905] env[70013]: DEBUG oslo_concurrency.lockutils [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] Acquiring lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.779061] env[70013]: DEBUG oslo_concurrency.lockutils [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] Acquired lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.779259] env[70013]: DEBUG nova.network.neutron [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Refreshing network info cache for port 022596bd-d9fb-42f4-b6ee-29eb5192e817 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 839.791620] env[70013]: DEBUG oslo_vmware.api [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146845} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.791896] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 839.792057] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 839.792243] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 839.792416] env[70013]: INFO nova.compute.manager [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Took 1.12 seconds to destroy the instance on the hypervisor. [ 839.792658] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 839.792848] env[70013]: DEBUG nova.compute.manager [-] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 839.792943] env[70013]: DEBUG nova.network.neutron [-] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 839.844339] env[70013]: DEBUG oslo_vmware.api [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230865, 'name': ReconfigVM_Task, 'duration_secs': 0.596372} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.845498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.845498] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Reconfigured VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 839.938159] env[70013]: DEBUG nova.compute.manager [req-d9bd7db3-8633-4368-8ca4-d44bbc13f745 req-bca520c8-025f-4c14-989f-6d035dddcc94 service nova] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Received event network-vif-deleted-addd72e9-8e61-463a-ab33-7872333bd242 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 839.938712] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32ad78c9-a7f6-42f6-aa8a-1e41601d6556 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.320s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.001696] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 840.136080] env[70013]: INFO nova.compute.manager [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Took 1.38 seconds to deallocate network for instance. [ 840.349566] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0089690f-23b0-4732-92f3-92077d806479 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-9a0ecd6c-5853-446d-8f41-b2ee51e5259a-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.582s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.507745] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 840.574173] env[70013]: DEBUG nova.network.neutron [-] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.610246] env[70013]: DEBUG nova.network.neutron [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updated VIF entry in instance network info cache for port 022596bd-d9fb-42f4-b6ee-29eb5192e817. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 840.610641] env[70013]: DEBUG nova.network.neutron [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "022596bd-d9fb-42f4-b6ee-29eb5192e817", "address": "fa:16:3e:ba:06:18", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022596bd-d9", "ovs_interfaceid": "022596bd-d9fb-42f4-b6ee-29eb5192e817", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.011446] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance eab0c393-4d3d-4659-8225-c82ddadb330b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 841.077621] env[70013]: INFO nova.compute.manager [-] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Took 1.28 seconds to deallocate network for instance. [ 841.115414] env[70013]: DEBUG oslo_concurrency.lockutils [req-8cada9fd-6e5b-4b47-b7dc-7074a882e6b8 req-366242b3-0dd0-47cd-afaa-8b2e0caac4d8 service nova] Releasing lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 841.170850] env[70013]: INFO nova.scheduler.client.report [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Deleted allocations for instance b1d844cc-dcdc-47e2-85c3-484d0862be60 [ 841.171129] env[70013]: DEBUG oslo_concurrency.lockutils [None req-564dc407-d78e-45ea-8729-ffc3de35dd31 tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.371s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.172506] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.589s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.172747] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "b1d844cc-dcdc-47e2-85c3-484d0862be60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.172956] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.173131] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.175668] env[70013]: INFO nova.compute.manager [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Terminating instance [ 841.516441] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance d84d8426-b3fd-4f0d-8410-0ee24fedb35d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 841.585237] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.675816] env[70013]: DEBUG nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 841.680712] env[70013]: DEBUG nova.compute.manager [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 841.680712] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 841.682045] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40643385-182e-454d-9857-880f41c9c17a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.689888] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 841.690176] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b9cbdf6-16c1-4cd5-87e8-2b100f6cb01f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.698613] env[70013]: DEBUG oslo_vmware.api [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 841.698613] env[70013]: value = "task-4230868" [ 841.698613] env[70013]: _type = "Task" [ 841.698613] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.707488] env[70013]: DEBUG oslo_vmware.api [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.018774] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 842.019114] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 842.019298] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4032MB phys_disk=200GB used_disk=17GB total_vcpus=48 used_vcpus=18 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 842.036714] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 842.052827] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 842.053039] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.061513] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "b853f0c4-b79e-4e30-8ba1-3d3fb7041809" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.061761] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "b853f0c4-b79e-4e30-8ba1-3d3fb7041809" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.069277] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 842.091196] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 842.144432] env[70013]: DEBUG nova.compute.manager [req-2c785249-711b-4756-9094-7e41309cdef6 req-aa6d5ae1-073b-41a7-8d05-5746e7b4c939 service nova] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Received event network-vif-deleted-6eb8e0b6-a88f-4ebb-b00f-2d3b86cf1230 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 842.198823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.208739] env[70013]: DEBUG oslo_vmware.api [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230868, 'name': PowerOffVM_Task, 'duration_secs': 0.201417} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.210966] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 842.211178] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 842.211625] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d733e39c-551f-43f9-836e-2146b447b9e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.280655] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 842.280902] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 842.281112] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Deleting the datastore file [datastore2] b1d844cc-dcdc-47e2-85c3-484d0862be60 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.281386] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd3ad039-a6d4-4e98-aa95-81f993fd930e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.289299] env[70013]: DEBUG oslo_vmware.api [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for the task: (returnval){ [ 842.289299] env[70013]: value = "task-4230870" [ 842.289299] env[70013]: _type = "Task" [ 842.289299] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.298699] env[70013]: DEBUG oslo_vmware.api [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230870, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.556854] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2727b8d-0504-4372-8f11-ec16bbbd05bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.565182] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8321c2f4-691e-471f-b209-b93656eb90e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.597502] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90119ad9-7e97-4d41-a751-aba567d7959d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.605908] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bff2b0c-a874-40f8-95ad-075cab9356a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.620224] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.754990] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-9a0ecd6c-5853-446d-8f41-b2ee51e5259a-022596bd-d9fb-42f4-b6ee-29eb5192e817" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.755289] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-9a0ecd6c-5853-446d-8f41-b2ee51e5259a-022596bd-d9fb-42f4-b6ee-29eb5192e817" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.811129] env[70013]: DEBUG oslo_vmware.api [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Task: {'id': task-4230870, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162447} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.811313] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.812092] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 842.812092] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 842.812092] env[70013]: INFO nova.compute.manager [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Took 1.13 seconds to destroy the instance on the hypervisor. [ 842.813547] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 842.813791] env[70013]: DEBUG nova.compute.manager [-] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 842.813873] env[70013]: DEBUG nova.network.neutron [-] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 843.128022] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 843.261020] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.261020] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.261020] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52ff8dc-600d-4fd9-8ade-ac6f26812275 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.287222] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4ddfcd-edf9-452f-8a46-0df5a3a43e8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.316069] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Reconfiguring VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 843.316635] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e8e9589-cd1d-4997-afe9-20e130c2e7ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.339667] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 843.339667] env[70013]: value = "task-4230871" [ 843.339667] env[70013]: _type = "Task" [ 843.339667] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.350539] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.353137] env[70013]: DEBUG nova.network.neutron [-] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.632481] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 843.632481] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.272s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.632481] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.646s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.634834] env[70013]: INFO nova.compute.claims [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.853616] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.856201] env[70013]: INFO nova.compute.manager [-] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Took 1.04 seconds to deallocate network for instance. [ 844.261199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Acquiring lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.261199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.351222] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.362930] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.763457] env[70013]: DEBUG nova.compute.utils [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 844.855872] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.122427] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719de649-1327-4116-821e-61234863b2c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.131661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cd78e9-9d42-4e00-afbf-63c8286b12e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.163108] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475d552c-bad7-40e9-a4f1-34c3470c93e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.171705] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8161db1-e53b-4592-83dd-29f7489ebc18 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.186190] env[70013]: DEBUG nova.compute.provider_tree [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.268092] env[70013]: DEBUG oslo_concurrency.lockutils [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.355998] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.692090] env[70013]: DEBUG nova.scheduler.client.report [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 845.861376] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.199121] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 846.199874] env[70013]: DEBUG nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 846.203027] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.146s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.204659] env[70013]: INFO nova.compute.claims [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.358054] env[70013]: DEBUG oslo_concurrency.lockutils [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Acquiring lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 846.358810] env[70013]: DEBUG oslo_concurrency.lockutils [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.359078] env[70013]: INFO nova.compute.manager [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Attaching volume 7c68bccf-7eb8-4bca-a0f2-8e81c644591e to /dev/sdb [ 846.362209] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.400017] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b711c9-0899-4f4c-890e-df55a2e6b745 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.406596] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b41b222-c267-4718-a907-f2db3d7423d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.422991] env[70013]: DEBUG nova.virt.block_device [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Updating existing volume attachment record: 7c040303-66d6-4dbb-92c4-5e43f5bffb27 {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 846.710540] env[70013]: DEBUG nova.compute.utils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 846.714839] env[70013]: DEBUG nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 846.717630] env[70013]: DEBUG nova.network.neutron [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.801468] env[70013]: DEBUG nova.policy [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfa9f992483942f689f616b60a27b00d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0896b76438d9456fb1b9fa2f39d0decc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 846.857775] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.216821] env[70013]: DEBUG nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 847.357629] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.624079] env[70013]: DEBUG nova.network.neutron [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Successfully created port: c0c18e7a-0f35-4d2e-b12d-0002001b6793 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.822847] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94bd858-6abd-415f-a376-da3d40f86b26 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.832111] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc28df4c-7d51-4d89-9d37-7929612f23fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.871424] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1d1d46-1010-4b02-91e0-9f7cf3cb7c51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.879500] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.883087] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e83899-8699-4367-9230-17576efb0f0c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.899469] env[70013]: DEBUG nova.compute.provider_tree [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.245487] env[70013]: DEBUG nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 848.275877] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 848.276150] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.276310] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 848.276495] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.276647] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 848.276818] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 848.277092] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 848.277262] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 848.277431] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 848.277643] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 848.277876] env[70013]: DEBUG nova.virt.hardware [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 848.278811] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1289ed8d-d72e-4180-ab2e-2000428918d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.288030] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08de2d41-cf2f-4c04-9be1-654d28979642 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.377286] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.403553] env[70013]: DEBUG nova.scheduler.client.report [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 848.879528] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.914023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.708s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.914023] env[70013]: DEBUG nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 848.916329] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.536s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 848.916867] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.919415] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.634s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 848.966440] env[70013]: INFO nova.scheduler.client.report [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Deleted allocations for instance 2ee62ad3-0125-47dc-b163-7d15b6a17c8a [ 849.258555] env[70013]: DEBUG nova.compute.manager [req-a8ff01bf-8f0d-4498-871c-6eb24792646e req-63524168-7c9f-42b1-b019-60d2bcfba459 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Received event network-vif-plugged-c0c18e7a-0f35-4d2e-b12d-0002001b6793 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 849.258839] env[70013]: DEBUG oslo_concurrency.lockutils [req-a8ff01bf-8f0d-4498-871c-6eb24792646e req-63524168-7c9f-42b1-b019-60d2bcfba459 service nova] Acquiring lock "f715cf45-81f1-4979-abfc-455b3e64f00e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.258971] env[70013]: DEBUG oslo_concurrency.lockutils [req-a8ff01bf-8f0d-4498-871c-6eb24792646e req-63524168-7c9f-42b1-b019-60d2bcfba459 service nova] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.259155] env[70013]: DEBUG oslo_concurrency.lockutils [req-a8ff01bf-8f0d-4498-871c-6eb24792646e req-63524168-7c9f-42b1-b019-60d2bcfba459 service nova] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.259391] env[70013]: DEBUG nova.compute.manager [req-a8ff01bf-8f0d-4498-871c-6eb24792646e req-63524168-7c9f-42b1-b019-60d2bcfba459 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] No waiting events found dispatching network-vif-plugged-c0c18e7a-0f35-4d2e-b12d-0002001b6793 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 849.259617] env[70013]: WARNING nova.compute.manager [req-a8ff01bf-8f0d-4498-871c-6eb24792646e req-63524168-7c9f-42b1-b019-60d2bcfba459 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Received unexpected event network-vif-plugged-c0c18e7a-0f35-4d2e-b12d-0002001b6793 for instance with vm_state building and task_state spawning. [ 849.377976] env[70013]: DEBUG oslo_vmware.api [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230871, 'name': ReconfigVM_Task, 'duration_secs': 5.782482} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.379036] env[70013]: DEBUG nova.network.neutron [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Successfully updated port: c0c18e7a-0f35-4d2e-b12d-0002001b6793 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.381896] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 849.381896] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Reconfigured VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 849.427364] env[70013]: DEBUG nova.compute.utils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 849.431819] env[70013]: DEBUG nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 849.432127] env[70013]: DEBUG nova.network.neutron [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 849.483114] env[70013]: DEBUG nova.policy [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac21e3c87f0d4c6fbff58c8bd524b181', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '719bc6e70189480dbef8b541a7a63d36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 849.490695] env[70013]: DEBUG oslo_concurrency.lockutils [None req-11e2379c-da57-46eb-8faa-c628c9000c16 tempest-ServersNegativeTestMultiTenantJSON-439672739 tempest-ServersNegativeTestMultiTenantJSON-439672739-project-member] Lock "2ee62ad3-0125-47dc-b163-7d15b6a17c8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.692s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.526832] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae535a3-ee17-435e-9df5-c848038be098 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.535729] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0137da7b-b526-4743-9720-e9de6b032785 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.580952] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f797ca-8ff4-47fd-83ef-b1a8ea7f8d3b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.589478] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4baedfb-58ed-4f71-b8c7-74d6716de1fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.614345] env[70013]: DEBUG nova.compute.provider_tree [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.884302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-f715cf45-81f1-4979-abfc-455b3e64f00e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.884899] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-f715cf45-81f1-4979-abfc-455b3e64f00e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.884899] env[70013]: DEBUG nova.network.neutron [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.905489] env[70013]: DEBUG nova.network.neutron [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Successfully created port: 84120f9b-ce4d-4d5a-9090-faca58cc4f6a {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.934970] env[70013]: DEBUG nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 850.119213] env[70013]: DEBUG nova.scheduler.client.report [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 850.121021] env[70013]: DEBUG oslo_concurrency.lockutils [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.201s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.121021] env[70013]: INFO nova.compute.manager [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Successfully reverted task state from None on failure for instance. [ 850.124095] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.963s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.127226] env[70013]: INFO nova.compute.claims [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server [None req-503df2ef-56ec-4211-ad1d-e3b293faaf8e tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server yield [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b71a728c-655c-4394-aca1-0ecf06dcab40"}]} [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 850.132538] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 850.133954] env[70013]: ERROR oslo_messaging.rpc.server [ 850.249987] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Acquiring lock "cfccb245-dec8-4dec-bd73-33fe18ec8a19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.250488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Lock "cfccb245-dec8-4dec-bd73-33fe18ec8a19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.417841] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.418140] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.418438] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.418732] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.418917] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.421239] env[70013]: INFO nova.compute.manager [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Terminating instance [ 850.424644] env[70013]: DEBUG nova.network.neutron [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.607096] env[70013]: DEBUG nova.network.neutron [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Updating instance_info_cache with network_info: [{"id": "c0c18e7a-0f35-4d2e-b12d-0002001b6793", "address": "fa:16:3e:cc:6f:b2", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0c18e7a-0f", "ovs_interfaceid": "c0c18e7a-0f35-4d2e-b12d-0002001b6793", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.756094] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.756295] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 850.756374] env[70013]: DEBUG nova.network.neutron [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.928034] env[70013]: DEBUG nova.compute.manager [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 850.928034] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.928034] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacf3e79-72e6-4d52-8a18-f2c4557f965d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.938572] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.939079] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e443f8d1-4caa-435c-bd0e-f6c104956312 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.951046] env[70013]: DEBUG nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 850.956043] env[70013]: DEBUG oslo_vmware.api [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 850.956043] env[70013]: value = "task-4230876" [ 850.956043] env[70013]: _type = "Task" [ 850.956043] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.966177] env[70013]: DEBUG oslo_vmware.api [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230876, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.990401] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 850.990783] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.990958] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 850.991370] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.991533] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 850.991681] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 850.991930] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 850.992111] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 850.992293] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 850.992459] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 850.992633] env[70013]: DEBUG nova.virt.hardware [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 850.993829] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9f5453-f40f-45ae-9612-5854f1c53ffe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.003508] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 851.003770] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837113', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'name': 'volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25', 'attached_at': '', 'detached_at': '', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'serial': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 851.005190] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e089b9-eeb7-4caf-8752-c940d4bee18e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.010618] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e22c3a7-cf7c-4d66-84e2-91babddf9039 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.049901] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd85908b-5e3f-4561-9bee-72ca6d3584a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.080869] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e/volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.081258] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05a8b02e-ffaa-42a1-8d97-07a3a32370e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.106336] env[70013]: DEBUG oslo_vmware.api [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Waiting for the task: (returnval){ [ 851.106336] env[70013]: value = "task-4230877" [ 851.106336] env[70013]: _type = "Task" [ 851.106336] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.114480] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-f715cf45-81f1-4979-abfc-455b3e64f00e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 851.114480] env[70013]: DEBUG nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Instance network_info: |[{"id": "c0c18e7a-0f35-4d2e-b12d-0002001b6793", "address": "fa:16:3e:cc:6f:b2", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0c18e7a-0f", "ovs_interfaceid": "c0c18e7a-0f35-4d2e-b12d-0002001b6793", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 851.119158] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:6f:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac7039c0-3374-4c08-87fc-af2449b48b02', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0c18e7a-0f35-4d2e-b12d-0002001b6793', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.128931] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 851.128931] env[70013]: DEBUG oslo_vmware.api [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230877, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.128931] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.128931] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d0534df1-7f76-411d-bb45-b585c426aaa4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.153278] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.153278] env[70013]: value = "task-4230878" [ 851.153278] env[70013]: _type = "Task" [ 851.153278] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.165200] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230878, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.425442] env[70013]: DEBUG nova.compute.manager [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Received event network-changed-c0c18e7a-0f35-4d2e-b12d-0002001b6793 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 851.425727] env[70013]: DEBUG nova.compute.manager [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Refreshing instance network info cache due to event network-changed-c0c18e7a-0f35-4d2e-b12d-0002001b6793. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 851.430023] env[70013]: DEBUG oslo_concurrency.lockutils [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] Acquiring lock "refresh_cache-f715cf45-81f1-4979-abfc-455b3e64f00e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.430023] env[70013]: DEBUG oslo_concurrency.lockutils [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] Acquired lock "refresh_cache-f715cf45-81f1-4979-abfc-455b3e64f00e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 851.430023] env[70013]: DEBUG nova.network.neutron [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Refreshing network info cache for port c0c18e7a-0f35-4d2e-b12d-0002001b6793 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.474498] env[70013]: DEBUG oslo_vmware.api [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230876, 'name': PowerOffVM_Task, 'duration_secs': 0.250702} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.474822] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.474994] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 851.475449] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-575a817a-0d5b-4b9a-83b5-e0049e69318f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.559358] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.559358] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.559358] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleting the datastore file [datastore1] 9a0ecd6c-5853-446d-8f41-b2ee51e5259a {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.559358] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6fe6c6c4-fad1-46af-8c0a-dee30bfe3d95 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.568941] env[70013]: DEBUG oslo_vmware.api [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 851.568941] env[70013]: value = "task-4230880" [ 851.568941] env[70013]: _type = "Task" [ 851.568941] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.581582] env[70013]: DEBUG oslo_vmware.api [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230880, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.619853] env[70013]: DEBUG oslo_vmware.api [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230877, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.661802] env[70013]: INFO nova.network.neutron [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Port 022596bd-d9fb-42f4-b6ee-29eb5192e817 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 851.662176] env[70013]: DEBUG nova.network.neutron [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.667791] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230878, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.693528] env[70013]: DEBUG nova.compute.manager [req-bb6625b7-5d44-4fd1-b37f-045ded7b43dc req-9492b281-228a-4ac6-8393-619f8c0a4267 service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Received event network-vif-plugged-84120f9b-ce4d-4d5a-9090-faca58cc4f6a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 851.693528] env[70013]: DEBUG oslo_concurrency.lockutils [req-bb6625b7-5d44-4fd1-b37f-045ded7b43dc req-9492b281-228a-4ac6-8393-619f8c0a4267 service nova] Acquiring lock "be9a565b-c74c-442e-8087-cc4579b1c8e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 851.693961] env[70013]: DEBUG oslo_concurrency.lockutils [req-bb6625b7-5d44-4fd1-b37f-045ded7b43dc req-9492b281-228a-4ac6-8393-619f8c0a4267 service nova] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 851.694193] env[70013]: DEBUG oslo_concurrency.lockutils [req-bb6625b7-5d44-4fd1-b37f-045ded7b43dc req-9492b281-228a-4ac6-8393-619f8c0a4267 service nova] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 851.694441] env[70013]: DEBUG nova.compute.manager [req-bb6625b7-5d44-4fd1-b37f-045ded7b43dc req-9492b281-228a-4ac6-8393-619f8c0a4267 service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] No waiting events found dispatching network-vif-plugged-84120f9b-ce4d-4d5a-9090-faca58cc4f6a {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 851.694656] env[70013]: WARNING nova.compute.manager [req-bb6625b7-5d44-4fd1-b37f-045ded7b43dc req-9492b281-228a-4ac6-8393-619f8c0a4267 service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Received unexpected event network-vif-plugged-84120f9b-ce4d-4d5a-9090-faca58cc4f6a for instance with vm_state building and task_state spawning. [ 851.833215] env[70013]: DEBUG nova.network.neutron [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Successfully updated port: 84120f9b-ce4d-4d5a-9090-faca58cc4f6a {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.839975] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795e51ba-6a73-4f23-87f5-18fdba7ba590 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.848973] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7357e351-0448-47bc-8898-23e45fc72207 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.886277] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ff07b8-6fb9-41eb-9c56-4c8623b8848a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.895116] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd0ced7-32f4-4fc6-be67-aa96af5c25ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.913406] env[70013]: DEBUG nova.compute.provider_tree [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.078714] env[70013]: DEBUG oslo_vmware.api [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4230880, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179937} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.078969] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.079168] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 852.079347] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 852.079550] env[70013]: INFO nova.compute.manager [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 852.079887] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 852.080111] env[70013]: DEBUG nova.compute.manager [-] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 852.080210] env[70013]: DEBUG nova.network.neutron [-] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 852.119669] env[70013]: DEBUG oslo_vmware.api [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230877, 'name': ReconfigVM_Task, 'duration_secs': 0.517975} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.120068] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Reconfigured VM instance instance-00000010 to attach disk [datastore2] volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e/volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.125710] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de9b38ab-2506-4219-b0fb-c7ee693dd06c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.143744] env[70013]: DEBUG oslo_vmware.api [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Waiting for the task: (returnval){ [ 852.143744] env[70013]: value = "task-4230881" [ 852.143744] env[70013]: _type = "Task" [ 852.143744] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.152483] env[70013]: DEBUG oslo_vmware.api [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.163415] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230878, 'name': CreateVM_Task, 'duration_secs': 0.873717} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.163578] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.164291] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.164442] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.164756] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 852.165258] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-9a0ecd6c-5853-446d-8f41-b2ee51e5259a" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 852.167559] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb0aad04-ced4-48c6-9ff7-6b9420d61f2e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.173704] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 852.173704] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b520ea-6327-dc65-e1d4-fdcf5688b675" [ 852.173704] env[70013]: _type = "Task" [ 852.173704] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.183324] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b520ea-6327-dc65-e1d4-fdcf5688b675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.185566] env[70013]: DEBUG nova.network.neutron [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Updated VIF entry in instance network info cache for port c0c18e7a-0f35-4d2e-b12d-0002001b6793. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.188695] env[70013]: DEBUG nova.network.neutron [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Updating instance_info_cache with network_info: [{"id": "c0c18e7a-0f35-4d2e-b12d-0002001b6793", "address": "fa:16:3e:cc:6f:b2", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0c18e7a-0f", "ovs_interfaceid": "c0c18e7a-0f35-4d2e-b12d-0002001b6793", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.342514] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "refresh_cache-be9a565b-c74c-442e-8087-cc4579b1c8e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.342514] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired lock "refresh_cache-be9a565b-c74c-442e-8087-cc4579b1c8e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.342514] env[70013]: DEBUG nova.network.neutron [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.417285] env[70013]: DEBUG nova.scheduler.client.report [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 852.655205] env[70013]: DEBUG oslo_vmware.api [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230881, 'name': ReconfigVM_Task, 'duration_secs': 0.258924} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.655696] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837113', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'name': 'volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25', 'attached_at': '', 'detached_at': '', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'serial': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 852.670748] env[70013]: DEBUG oslo_concurrency.lockutils [None req-33ee1a00-f4c4-4335-94fa-cf0331b5eff2 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-9a0ecd6c-5853-446d-8f41-b2ee51e5259a-022596bd-d9fb-42f4-b6ee-29eb5192e817" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.915s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.686911] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b520ea-6327-dc65-e1d4-fdcf5688b675, 'name': SearchDatastore_Task, 'duration_secs': 0.010342} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.687242] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 852.687486] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.687764] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.688184] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.688184] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.689282] env[70013]: DEBUG oslo_concurrency.lockutils [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] Releasing lock "refresh_cache-f715cf45-81f1-4979-abfc-455b3e64f00e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 852.689511] env[70013]: DEBUG nova.compute.manager [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received event network-vif-deleted-022596bd-d9fb-42f4-b6ee-29eb5192e817 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 852.689768] env[70013]: INFO nova.compute.manager [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Neutron deleted interface 022596bd-d9fb-42f4-b6ee-29eb5192e817; detaching it from the instance and deleting it from the info cache [ 852.690056] env[70013]: DEBUG nova.network.neutron [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [{"id": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "address": "fa:16:3e:e8:6e:81", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddab8d9-c5", "ovs_interfaceid": "3ddab8d9-c5bd-4de2-994d-621e5e86cc88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.691588] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51c31c2e-cd33-4744-b46d-0433ab9fb4e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.705894] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.705894] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.705894] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5ebfcfb-3d80-456b-9432-877de4ff6c0c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.716777] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 852.716777] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521dbcac-defe-824f-ffe2-ec3caaec31f7" [ 852.716777] env[70013]: _type = "Task" [ 852.716777] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.728401] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521dbcac-defe-824f-ffe2-ec3caaec31f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.883316] env[70013]: DEBUG nova.network.neutron [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.923691] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.799s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.924266] env[70013]: DEBUG nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 852.926871] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.735s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.930080] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.930080] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.330s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.931254] env[70013]: INFO nova.compute.claims [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.975851] env[70013]: INFO nova.scheduler.client.report [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Deleted allocations for instance ac8aa41c-375a-4231-9a5d-12d6c4bba44c [ 853.060276] env[70013]: DEBUG nova.network.neutron [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Updating instance_info_cache with network_info: [{"id": "84120f9b-ce4d-4d5a-9090-faca58cc4f6a", "address": "fa:16:3e:8c:fa:a7", "network": {"id": "1c5012af-830c-4001-9013-9e8cf32698bb", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-222945680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "719bc6e70189480dbef8b541a7a63d36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84120f9b-ce", "ovs_interfaceid": "84120f9b-ce4d-4d5a-9090-faca58cc4f6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.195229] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c89080b-2d31-4302-aaf8-a98f2b02c3f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.219396] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d60b0f-fe44-478d-bdc5-7e0af05df81a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.236107] env[70013]: DEBUG nova.objects.instance [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lazy-loading 'flavor' on Instance uuid b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.250480] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521dbcac-defe-824f-ffe2-ec3caaec31f7, 'name': SearchDatastore_Task, 'duration_secs': 0.010452} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.251339] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c393cf9-646b-4fe3-a31d-adf591b9a00f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.268443] env[70013]: DEBUG nova.compute.manager [req-e537468f-8b46-4822-98a3-488c086c398b req-78d54357-fb32-4b62-9194-fb2b705a0ae0 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Detach interface failed, port_id=022596bd-d9fb-42f4-b6ee-29eb5192e817, reason: Instance 9a0ecd6c-5853-446d-8f41-b2ee51e5259a could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 853.269778] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 853.269778] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52034e75-3f7c-55ab-07e0-a5bf516b9288" [ 853.269778] env[70013]: _type = "Task" [ 853.269778] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.280179] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52034e75-3f7c-55ab-07e0-a5bf516b9288, 'name': SearchDatastore_Task, 'duration_secs': 0.013431} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.280451] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 853.280812] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f715cf45-81f1-4979-abfc-455b3e64f00e/f715cf45-81f1-4979-abfc-455b3e64f00e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.281144] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52801ec5-b2a6-44b2-8960-3e650bb8dd87 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.290682] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 853.290682] env[70013]: value = "task-4230882" [ 853.290682] env[70013]: _type = "Task" [ 853.290682] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.301543] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.436397] env[70013]: DEBUG nova.compute.utils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 853.438311] env[70013]: DEBUG nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 853.438311] env[70013]: DEBUG nova.network.neutron [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.488366] env[70013]: DEBUG oslo_concurrency.lockutils [None req-92b7de8a-5dd4-4f7b-be6a-79fd99922855 tempest-ServersAaction247Test-1867371537 tempest-ServersAaction247Test-1867371537-project-member] Lock "ac8aa41c-375a-4231-9a5d-12d6c4bba44c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.556s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.511392] env[70013]: DEBUG nova.policy [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '40c2f161f08c4338aee62a87db16b0c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a47ea4715673458a82eedca76b444180', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 853.564081] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Releasing lock "refresh_cache-be9a565b-c74c-442e-8087-cc4579b1c8e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 853.564664] env[70013]: DEBUG nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Instance network_info: |[{"id": "84120f9b-ce4d-4d5a-9090-faca58cc4f6a", "address": "fa:16:3e:8c:fa:a7", "network": {"id": "1c5012af-830c-4001-9013-9e8cf32698bb", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-222945680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "719bc6e70189480dbef8b541a7a63d36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84120f9b-ce", "ovs_interfaceid": "84120f9b-ce4d-4d5a-9090-faca58cc4f6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 853.565368] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:fa:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24b91e3b-50f1-4a16-b929-942e6b31b2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84120f9b-ce4d-4d5a-9090-faca58cc4f6a', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.580220] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 853.580220] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.580220] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6a83b79-7eef-4ff3-9e1f-b0c8c7d6c74d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.608252] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.608252] env[70013]: value = "task-4230883" [ 853.608252] env[70013]: _type = "Task" [ 853.608252] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.620364] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230883, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.742294] env[70013]: DEBUG oslo_concurrency.lockutils [None req-117f2c76-9cc2-4dc6-b132-8a3cdef3b238 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.384s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.787229] env[70013]: DEBUG nova.compute.manager [req-ca3111f4-75c7-4c4a-9a83-21fdcf32ff87 req-27e8b7a6-2bc5-4bf8-a230-76a35ad6d0d4 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Received event network-vif-deleted-3ddab8d9-c5bd-4de2-994d-621e5e86cc88 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 853.787229] env[70013]: INFO nova.compute.manager [req-ca3111f4-75c7-4c4a-9a83-21fdcf32ff87 req-27e8b7a6-2bc5-4bf8-a230-76a35ad6d0d4 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Neutron deleted interface 3ddab8d9-c5bd-4de2-994d-621e5e86cc88; detaching it from the instance and deleting it from the info cache [ 853.787229] env[70013]: DEBUG nova.network.neutron [req-ca3111f4-75c7-4c4a-9a83-21fdcf32ff87 req-27e8b7a6-2bc5-4bf8-a230-76a35ad6d0d4 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.807402] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230882, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.911668] env[70013]: DEBUG nova.compute.manager [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Received event network-changed-84120f9b-ce4d-4d5a-9090-faca58cc4f6a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 853.911949] env[70013]: DEBUG nova.compute.manager [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Refreshing instance network info cache due to event network-changed-84120f9b-ce4d-4d5a-9090-faca58cc4f6a. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 853.914333] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] Acquiring lock "refresh_cache-be9a565b-c74c-442e-8087-cc4579b1c8e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.914333] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] Acquired lock "refresh_cache-be9a565b-c74c-442e-8087-cc4579b1c8e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 853.914333] env[70013]: DEBUG nova.network.neutron [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Refreshing network info cache for port 84120f9b-ce4d-4d5a-9090-faca58cc4f6a {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.944656] env[70013]: DEBUG nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 854.036511] env[70013]: DEBUG nova.network.neutron [-] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.122896] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230883, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.178721] env[70013]: DEBUG nova.network.neutron [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Successfully created port: 3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.296017] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94803d10-f5b0-47ac-8db8-8edadbdb4b5f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.316976] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230882, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629801} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.319278] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f715cf45-81f1-4979-abfc-455b3e64f00e/f715cf45-81f1-4979-abfc-455b3e64f00e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.319533] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.322876] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7060586f-bb88-4947-93cf-1bc46e75166a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.332525] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effe82bb-0a95-46d3-9294-f0312b9ad4a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.354348] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 854.354348] env[70013]: value = "task-4230884" [ 854.354348] env[70013]: _type = "Task" [ 854.354348] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.378801] env[70013]: DEBUG nova.compute.manager [req-ca3111f4-75c7-4c4a-9a83-21fdcf32ff87 req-27e8b7a6-2bc5-4bf8-a230-76a35ad6d0d4 service nova] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Detach interface failed, port_id=3ddab8d9-c5bd-4de2-994d-621e5e86cc88, reason: Instance 9a0ecd6c-5853-446d-8f41-b2ee51e5259a could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 854.389870] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230884, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.543349] env[70013]: INFO nova.compute.manager [-] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Took 2.46 seconds to deallocate network for instance. [ 854.632150] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230883, 'name': CreateVM_Task, 'duration_secs': 0.859766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.632150] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.632770] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.633202] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.633595] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 854.634362] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b36587f-df4a-4b9c-a70d-fdd448a08888 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.647080] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 854.647080] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522dc84c-d982-49cc-8d04-1b1d0d5112d4" [ 854.647080] env[70013]: _type = "Task" [ 854.647080] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.661781] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522dc84c-d982-49cc-8d04-1b1d0d5112d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.696503] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f0c05f-6287-465e-a113-ac9e575781d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.704996] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871bd606-3b8d-44d8-9f36-e74f20f21976 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.740045] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e54017-c81e-40cb-8ea1-19eda68b5d1b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.747801] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c045a477-0e52-4fd1-b9c4-27a7873e5e42 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.762735] env[70013]: DEBUG nova.compute.provider_tree [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 854.866801] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230884, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103741} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.868073] env[70013]: DEBUG nova.network.neutron [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Updated VIF entry in instance network info cache for port 84120f9b-ce4d-4d5a-9090-faca58cc4f6a. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 854.868423] env[70013]: DEBUG nova.network.neutron [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Updating instance_info_cache with network_info: [{"id": "84120f9b-ce4d-4d5a-9090-faca58cc4f6a", "address": "fa:16:3e:8c:fa:a7", "network": {"id": "1c5012af-830c-4001-9013-9e8cf32698bb", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-222945680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "719bc6e70189480dbef8b541a7a63d36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84120f9b-ce", "ovs_interfaceid": "84120f9b-ce4d-4d5a-9090-faca58cc4f6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.869981] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.870776] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289ab493-baa0-44bf-8bc4-aee17fc11a2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.899879] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] f715cf45-81f1-4979-abfc-455b3e64f00e/f715cf45-81f1-4979-abfc-455b3e64f00e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.900636] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84ebfb22-7061-4df9-98b2-cfc94c0e5bde {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.927992] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 854.927992] env[70013]: value = "task-4230885" [ 854.927992] env[70013]: _type = "Task" [ 854.927992] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.939345] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230885, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.964440] env[70013]: DEBUG nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 854.991634] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 854.991858] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.992115] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 854.992315] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.992463] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 854.992646] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 854.992902] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 854.994086] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 854.994317] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 854.994534] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 854.994687] env[70013]: DEBUG nova.virt.hardware [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 854.995941] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4547763a-84be-4f51-9c38-6b509d80b30e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.004873] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54529b53-b715-4570-93de-a2e3411c0cc8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.056180] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.160778] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522dc84c-d982-49cc-8d04-1b1d0d5112d4, 'name': SearchDatastore_Task, 'duration_secs': 0.041893} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.161126] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.161287] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.161553] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.161702] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.161884] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.162468] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5741969d-225b-4d4b-8f76-6bd60d164264 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.175430] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.175627] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.176466] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2e84f01-7b25-45db-877a-e1b14eb35128 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.188781] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 855.188781] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab13d2-d2b5-dbe7-4b87-afb5dd93a2a5" [ 855.188781] env[70013]: _type = "Task" [ 855.188781] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.199339] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab13d2-d2b5-dbe7-4b87-afb5dd93a2a5, 'name': SearchDatastore_Task, 'duration_secs': 0.012946} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.200857] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2655fd69-c44a-4e04-bc90-26ab6a6db0b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.206125] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 855.206125] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200c1fc-0ccf-b2c2-8cb6-1bc3da798f73" [ 855.206125] env[70013]: _type = "Task" [ 855.206125] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.218151] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200c1fc-0ccf-b2c2-8cb6-1bc3da798f73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.286228] env[70013]: ERROR nova.scheduler.client.report [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [req-f943587b-493c-4696-acf1-e955c9257b7b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f943587b-493c-4696-acf1-e955c9257b7b"}]} [ 855.287860] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 855.287860] env[70013]: ERROR nova.compute.manager [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Traceback (most recent call last): [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] yield [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] self.set_inventory_for_provider( [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f943587b-493c-4696-acf1-e955c9257b7b"}]} [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] During handling of the above exception, another exception occurred: [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Traceback (most recent call last): [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] with self.rt.instance_claim(context, instance, node, allocs, [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] return f(*args, **kwargs) [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] self._update(elevated, cn) [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] self._update_to_placement(context, compute_node, startup) [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] return attempt.get(self._wrap_exception) [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] six.reraise(self.value[0], self.value[1], self.value[2]) [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] raise value [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 855.287860] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] self.reportclient.update_from_provider_tree( [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] with catch_all(pd.uuid): [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] self.gen.throw(typ, value, traceback) [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] raise exception.ResourceProviderSyncFailed() [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 855.288982] env[70013]: ERROR nova.compute.manager [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] [ 855.288982] env[70013]: DEBUG nova.compute.utils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 855.291925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.053s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.291925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 855.295235] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.554s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.295855] env[70013]: INFO nova.compute.claims [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.299665] env[70013]: DEBUG nova.compute.manager [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Build of instance 57485f78-b97e-423c-b1f1-8648e963e5dd was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 855.300182] env[70013]: DEBUG nova.compute.manager [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 855.300453] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-57485f78-b97e-423c-b1f1-8648e963e5dd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.300640] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-57485f78-b97e-423c-b1f1-8648e963e5dd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.300993] env[70013]: DEBUG nova.network.neutron [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.325943] env[70013]: INFO nova.scheduler.client.report [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Deleted allocations for instance 3e938bb3-54a6-49f8-809d-d5aee0349eae [ 855.374712] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f6411e5-8756-4d1c-b986-4baae13b9059 req-30251a56-9ca8-4930-8ff6-8ccc32ed6f1c service nova] Releasing lock "refresh_cache-be9a565b-c74c-442e-8087-cc4579b1c8e7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.443019] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230885, 'name': ReconfigVM_Task, 'duration_secs': 0.349154} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.443384] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Reconfigured VM instance instance-0000002a to attach disk [datastore2] f715cf45-81f1-4979-abfc-455b3e64f00e/f715cf45-81f1-4979-abfc-455b3e64f00e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.444333] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4bc5fa2e-69fd-450e-8958-1d038319c50d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.452158] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 855.452158] env[70013]: value = "task-4230886" [ 855.452158] env[70013]: _type = "Task" [ 855.452158] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.462778] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230886, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.690909] env[70013]: INFO nova.compute.manager [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Rebuilding instance [ 855.721885] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200c1fc-0ccf-b2c2-8cb6-1bc3da798f73, 'name': SearchDatastore_Task, 'duration_secs': 0.018712} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.725489] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.725489] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] be9a565b-c74c-442e-8087-cc4579b1c8e7/be9a565b-c74c-442e-8087-cc4579b1c8e7.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.726972] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-85ea042b-f339-4bd3-9511-d2b0faf415bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.735810] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 855.735810] env[70013]: value = "task-4230887" [ 855.735810] env[70013]: _type = "Task" [ 855.735810] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.748392] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230887, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.751578] env[70013]: DEBUG nova.compute.manager [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 855.752453] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e6657a-720c-405c-be2f-b25700dade9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.834495] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e2ee1783-f5c8-480b-8d04-fe71677e6d6c tempest-ServersAdminNegativeTestJSON-1617245453 tempest-ServersAdminNegativeTestJSON-1617245453-project-member] Lock "3e938bb3-54a6-49f8-809d-d5aee0349eae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.036s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 855.839519] env[70013]: DEBUG nova.network.neutron [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.919171] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "965480de-4fe2-48ce-94ad-0838d78239dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.919481] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "965480de-4fe2-48ce-94ad-0838d78239dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.967804] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230886, 'name': Rename_Task, 'duration_secs': 0.231171} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.968362] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.968723] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b2acc75-17b5-4351-a040-57f7a7581ade {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.974165] env[70013]: DEBUG nova.network.neutron [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.981557] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 855.981557] env[70013]: value = "task-4230888" [ 855.981557] env[70013]: _type = "Task" [ 855.981557] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.997331] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.116839] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Acquiring lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.116976] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.158542] env[70013]: DEBUG nova.network.neutron [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Successfully updated port: 3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.204993] env[70013]: DEBUG nova.compute.manager [req-7b5b7b43-19e3-460c-904f-edcc077deeac req-5aee618c-9c84-4505-a240-353f3b2b07bf service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Received event network-vif-plugged-3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 856.204993] env[70013]: DEBUG oslo_concurrency.lockutils [req-7b5b7b43-19e3-460c-904f-edcc077deeac req-5aee618c-9c84-4505-a240-353f3b2b07bf service nova] Acquiring lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.204993] env[70013]: DEBUG oslo_concurrency.lockutils [req-7b5b7b43-19e3-460c-904f-edcc077deeac req-5aee618c-9c84-4505-a240-353f3b2b07bf service nova] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.204993] env[70013]: DEBUG oslo_concurrency.lockutils [req-7b5b7b43-19e3-460c-904f-edcc077deeac req-5aee618c-9c84-4505-a240-353f3b2b07bf service nova] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.204993] env[70013]: DEBUG nova.compute.manager [req-7b5b7b43-19e3-460c-904f-edcc077deeac req-5aee618c-9c84-4505-a240-353f3b2b07bf service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] No waiting events found dispatching network-vif-plugged-3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 856.205814] env[70013]: WARNING nova.compute.manager [req-7b5b7b43-19e3-460c-904f-edcc077deeac req-5aee618c-9c84-4505-a240-353f3b2b07bf service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Received unexpected event network-vif-plugged-3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 for instance with vm_state building and task_state spawning. [ 856.252894] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230887, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506465} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.253230] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] be9a565b-c74c-442e-8087-cc4579b1c8e7/be9a565b-c74c-442e-8087-cc4579b1c8e7.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.253489] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.253787] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-790f73af-1feb-4a2c-a73d-1d9bd6e41124 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.261955] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 856.261955] env[70013]: value = "task-4230889" [ 856.261955] env[70013]: _type = "Task" [ 856.261955] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.276952] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230889, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.334250] env[70013]: DEBUG nova.scheduler.client.report [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 856.353171] env[70013]: DEBUG nova.scheduler.client.report [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 856.353663] env[70013]: DEBUG nova.compute.provider_tree [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 856.366523] env[70013]: DEBUG nova.scheduler.client.report [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 856.389769] env[70013]: DEBUG nova.scheduler.client.report [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 856.477530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-57485f78-b97e-423c-b1f1-8648e963e5dd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.477843] env[70013]: DEBUG nova.compute.manager [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 856.478084] env[70013]: DEBUG nova.compute.manager [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 856.478368] env[70013]: DEBUG nova.network.neutron [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 856.496365] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230888, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.508755] env[70013]: DEBUG nova.network.neutron [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.620462] env[70013]: INFO nova.compute.manager [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Detaching volume 7c68bccf-7eb8-4bca-a0f2-8e81c644591e [ 856.664329] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.664443] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquired lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 856.664615] env[70013]: DEBUG nova.network.neutron [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.674321] env[70013]: INFO nova.virt.block_device [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Attempting to driver detach volume 7c68bccf-7eb8-4bca-a0f2-8e81c644591e from mountpoint /dev/sdb [ 856.674528] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 856.674646] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837113', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'name': 'volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25', 'attached_at': '', 'detached_at': '', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'serial': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 856.675884] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bea1a26-b9b9-48ce-a864-19d76c14e422 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.705311] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b50934-7387-4054-b9c1-5672e4d382f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.714585] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6991a8a-fc1f-4651-94d5-cfb13bc3247c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.740977] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d13cffe-5736-4e76-93e2-4168d3923037 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.763110] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] The volume has not been displaced from its original location: [datastore2] volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e/volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 856.769149] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Reconfiguring VM instance instance-00000010 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 856.773206] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 856.774103] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-566b42b2-66c3-47c6-833c-cc443f0a6068 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.791734] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7ce7c79-dac0-49bb-b14c-436325f5f1e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.799389] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230889, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075017} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.801194] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.801194] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 856.801194] env[70013]: value = "task-4230891" [ 856.801194] env[70013]: _type = "Task" [ 856.801194] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.802815] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9e8a5d-254c-4317-975a-bd9bf255d18b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.805945] env[70013]: DEBUG oslo_vmware.api [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Waiting for the task: (returnval){ [ 856.805945] env[70013]: value = "task-4230890" [ 856.805945] env[70013]: _type = "Task" [ 856.805945] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.837128] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] be9a565b-c74c-442e-8087-cc4579b1c8e7/be9a565b-c74c-442e-8087-cc4579b1c8e7.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.843609] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b8f0ebe-3617-4c6b-8efd-b791f1e9fa25 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.858161] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230891, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.862946] env[70013]: DEBUG oslo_vmware.api [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.871433] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 856.871433] env[70013]: value = "task-4230892" [ 856.871433] env[70013]: _type = "Task" [ 856.871433] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.878314] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.878314] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.878314] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.878314] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.878986] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.890085] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230892, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.891088] env[70013]: INFO nova.compute.manager [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Terminating instance [ 856.996588] env[70013]: DEBUG oslo_vmware.api [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230888, 'name': PowerOnVM_Task, 'duration_secs': 0.601888} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.999686] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.999926] env[70013]: INFO nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Took 8.75 seconds to spawn the instance on the hypervisor. [ 857.000152] env[70013]: DEBUG nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 857.001305] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03745ae1-da21-4e9f-a062-5ea6c244855d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.010733] env[70013]: DEBUG nova.network.neutron [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.121051] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fdffa7-9a81-4757-b9b3-7cac5dd4b758 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.129904] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c02fb4a-418f-4288-bc55-f3f0ee1dc8ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.165829] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef265b8-2f86-4985-b4e3-5639415bddbc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.182343] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc61d69b-adf4-4801-b45d-6fbd21c3c8b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.197075] env[70013]: DEBUG nova.compute.provider_tree [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 857.211492] env[70013]: DEBUG nova.network.neutron [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.315693] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230891, 'name': PowerOffVM_Task, 'duration_secs': 0.266495} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.318837] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.319561] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.319888] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f35c4af8-7177-411d-a14a-406ab14c2a9a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.327671] env[70013]: DEBUG oslo_vmware.api [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230890, 'name': ReconfigVM_Task, 'duration_secs': 0.346858} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.329398] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Reconfigured VM instance instance-00000010 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 857.334541] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 857.334541] env[70013]: value = "task-4230893" [ 857.334541] env[70013]: _type = "Task" [ 857.334541] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.334820] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6a27397-9946-46e3-b2f8-a4c15450f7cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.358929] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 857.359177] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 857.359373] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837062', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'name': 'volume-19da385b-04e0-4e04-9b4d-39b9370bec62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5c5fbdae-4916-4beb-ada0-57f36fb0f84c', 'attached_at': '', 'detached_at': '', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'serial': '19da385b-04e0-4e04-9b4d-39b9370bec62'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 857.359752] env[70013]: DEBUG oslo_vmware.api [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Waiting for the task: (returnval){ [ 857.359752] env[70013]: value = "task-4230894" [ 857.359752] env[70013]: _type = "Task" [ 857.359752] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.360486] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef42519c-1d2d-4ee8-b32c-a5af98eb50eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.372712] env[70013]: DEBUG oslo_vmware.api [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230894, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.390446] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcfefc1-7ef5-4d8f-9da4-d37d16792f02 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.393875] env[70013]: DEBUG nova.network.neutron [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Updating instance_info_cache with network_info: [{"id": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "address": "fa:16:3e:1d:87:e4", "network": {"id": "a3991650-c4b9-4370-9a29-d0ad4de0877f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-440150710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47ea4715673458a82eedca76b444180", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0bd58d-e1", "ovs_interfaceid": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.404320] env[70013]: DEBUG nova.compute.manager [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 857.404534] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 857.404832] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230892, 'name': ReconfigVM_Task, 'duration_secs': 0.433654} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.405672] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b922f87-1d53-48ee-9845-5ea853fc738d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.408337] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Reconfigured VM instance instance-0000002b to attach disk [datastore2] be9a565b-c74c-442e-8087-cc4579b1c8e7/be9a565b-c74c-442e-8087-cc4579b1c8e7.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.409412] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a627cda8-5b60-4eab-9be2-70492b38bfa2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.411858] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12297fee-f338-4151-9247-bfab1f23e6e1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.435486] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8854b76b-bd77-46b1-b3ce-1a846588d409 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.438408] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 857.438408] env[70013]: value = "task-4230895" [ 857.438408] env[70013]: _type = "Task" [ 857.438408] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.438609] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.439121] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6841cf7-e2f9-4f06-a357-6ab32fc59cc2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.455943] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] The volume has not been displaced from its original location: [datastore1] volume-19da385b-04e0-4e04-9b4d-39b9370bec62/volume-19da385b-04e0-4e04-9b4d-39b9370bec62.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 857.461305] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Reconfiguring VM instance instance-00000021 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 857.463441] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37c97eaf-7953-47b4-836d-f8fc48de9ea9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.479364] env[70013]: DEBUG oslo_vmware.api [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 857.479364] env[70013]: value = "task-4230896" [ 857.479364] env[70013]: _type = "Task" [ 857.479364] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.482777] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230895, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.489363] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 857.489363] env[70013]: value = "task-4230897" [ 857.489363] env[70013]: _type = "Task" [ 857.489363] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.493232] env[70013]: DEBUG oslo_vmware.api [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230896, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.503246] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230897, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.518871] env[70013]: INFO nova.compute.manager [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 57485f78-b97e-423c-b1f1-8648e963e5dd] Took 1.04 seconds to deallocate network for instance. [ 857.528314] env[70013]: INFO nova.compute.manager [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Took 45.58 seconds to build instance. [ 857.725544] env[70013]: ERROR nova.scheduler.client.report [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [req-0bfc6d2d-54ee-4346-b1a1-dc5d564115a0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0bfc6d2d-54ee-4346-b1a1-dc5d564115a0"}]} [ 857.725943] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.726634] env[70013]: ERROR nova.compute.manager [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Traceback (most recent call last): [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] yield [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] self.set_inventory_for_provider( [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0bfc6d2d-54ee-4346-b1a1-dc5d564115a0"}]} [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] During handling of the above exception, another exception occurred: [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Traceback (most recent call last): [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] with self.rt.instance_claim(context, instance, node, allocs, [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] return f(*args, **kwargs) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] self._update(elevated, cn) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] self._update_to_placement(context, compute_node, startup) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] return attempt.get(self._wrap_exception) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] six.reraise(self.value[0], self.value[1], self.value[2]) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] raise value [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] self.reportclient.update_from_provider_tree( [ 857.726634] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 857.728083] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] with catch_all(pd.uuid): [ 857.728083] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 857.728083] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] self.gen.throw(typ, value, traceback) [ 857.728083] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 857.728083] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] raise exception.ResourceProviderSyncFailed() [ 857.728083] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 857.728083] env[70013]: ERROR nova.compute.manager [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] [ 857.728083] env[70013]: DEBUG nova.compute.utils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 857.729099] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.954s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.730183] env[70013]: INFO nova.compute.claims [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.732894] env[70013]: DEBUG nova.compute.manager [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Build of instance 477bf0d8-e988-43e9-a573-994a0d08df0c was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 857.733332] env[70013]: DEBUG nova.compute.manager [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 857.733569] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "refresh_cache-477bf0d8-e988-43e9-a573-994a0d08df0c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.733695] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquired lock "refresh_cache-477bf0d8-e988-43e9-a573-994a0d08df0c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 857.733855] env[70013]: DEBUG nova.network.neutron [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.873941] env[70013]: DEBUG oslo_vmware.api [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Task: {'id': task-4230894, 'name': ReconfigVM_Task, 'duration_secs': 0.158365} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.874272] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837113', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'name': 'volume-7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25', 'attached_at': '', 'detached_at': '', 'volume_id': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e', 'serial': '7c68bccf-7eb8-4bca-a0f2-8e81c644591e'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 857.896746] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Releasing lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 857.897150] env[70013]: DEBUG nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Instance network_info: |[{"id": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "address": "fa:16:3e:1d:87:e4", "network": {"id": "a3991650-c4b9-4370-9a29-d0ad4de0877f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-440150710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47ea4715673458a82eedca76b444180", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0bd58d-e1", "ovs_interfaceid": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 857.897654] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:87:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.905501] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Creating folder: Project (a47ea4715673458a82eedca76b444180). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.905802] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d426dc65-d662-4026-96fd-6b9f7947e61f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.917478] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Created folder: Project (a47ea4715673458a82eedca76b444180) in parent group-v836999. [ 857.917686] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Creating folder: Instances. Parent ref: group-v837116. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.918237] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f32fcd6-558a-4bcd-a7d9-ccb64c8eeebe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.930075] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Created folder: Instances in parent group-v837116. [ 857.930355] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 857.930561] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.930796] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e315117f-9d99-49a2-94ec-515e6332ea97 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.955999] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230895, 'name': Rename_Task, 'duration_secs': 0.17994} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.958213] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.958213] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.958213] env[70013]: value = "task-4230900" [ 857.958213] env[70013]: _type = "Task" [ 857.958213] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.958213] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27f16eaa-6905-4620-99d8-40080c1d141e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.970684] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230900, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.972386] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 857.972386] env[70013]: value = "task-4230901" [ 857.972386] env[70013]: _type = "Task" [ 857.972386] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.981942] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.993830] env[70013]: DEBUG oslo_vmware.api [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230896, 'name': PowerOffVM_Task, 'duration_secs': 0.254095} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.997581] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.997813] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 857.998188] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-413d8e40-91fa-4c18-964d-bd04e5536f13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.007876] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230897, 'name': ReconfigVM_Task, 'duration_secs': 0.200282} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.008435] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Reconfigured VM instance instance-00000021 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 858.015655] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0019870e-8c0a-467d-b3fc-29cdf7a6ba56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.038267] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cd51f2f5-2a44-4e15-8b5d-03f66db98006 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.422s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 858.043726] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 858.043726] env[70013]: value = "task-4230903" [ 858.043726] env[70013]: _type = "Task" [ 858.043726] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.054653] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230903, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.077935] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.078588] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.078588] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Deleting the datastore file [datastore2] b23b4e67-67a6-4117-9a62-3c0dcd25e828 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.079195] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba27b2b1-6dee-44ba-98b3-d599ca33453d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.089165] env[70013]: DEBUG oslo_vmware.api [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for the task: (returnval){ [ 858.089165] env[70013]: value = "task-4230904" [ 858.089165] env[70013]: _type = "Task" [ 858.089165] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.098506] env[70013]: DEBUG oslo_vmware.api [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.246653] env[70013]: DEBUG nova.compute.manager [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Received event network-changed-3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 858.246875] env[70013]: DEBUG nova.compute.manager [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Refreshing instance network info cache due to event network-changed-3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 858.247113] env[70013]: DEBUG oslo_concurrency.lockutils [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] Acquiring lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.247294] env[70013]: DEBUG oslo_concurrency.lockutils [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] Acquired lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.247483] env[70013]: DEBUG nova.network.neutron [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Refreshing network info cache for port 3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.264194] env[70013]: DEBUG nova.network.neutron [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.381496] env[70013]: DEBUG nova.network.neutron [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.426802] env[70013]: DEBUG nova.objects.instance [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lazy-loading 'flavor' on Instance uuid b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.473708] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230900, 'name': CreateVM_Task, 'duration_secs': 0.41744} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.478596] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.479493] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.479762] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.480291] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 858.481126] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a25ab7dd-6c24-43a8-9919-090c0ce8fa0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.487972] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230901, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.492189] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 858.492189] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523ffddf-b2a2-f357-7eb9-31cd4fa607ea" [ 858.492189] env[70013]: _type = "Task" [ 858.492189] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.502775] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523ffddf-b2a2-f357-7eb9-31cd4fa607ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.541231] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "f715cf45-81f1-4979-abfc-455b3e64f00e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 858.541589] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 858.541589] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "f715cf45-81f1-4979-abfc-455b3e64f00e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 858.541705] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 858.541945] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 858.543639] env[70013]: DEBUG nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 858.546945] env[70013]: INFO nova.compute.manager [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Terminating instance [ 858.562480] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230903, 'name': ReconfigVM_Task, 'duration_secs': 0.160794} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.562851] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837062', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'name': 'volume-19da385b-04e0-4e04-9b4d-39b9370bec62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5c5fbdae-4916-4beb-ada0-57f36fb0f84c', 'attached_at': '', 'detached_at': '', 'volume_id': '19da385b-04e0-4e04-9b4d-39b9370bec62', 'serial': '19da385b-04e0-4e04-9b4d-39b9370bec62'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 858.563303] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 858.564066] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17aa61be-1934-4d0b-9e8b-8af0e40df9e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.573553] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 858.574588] env[70013]: INFO nova.scheduler.client.report [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted allocations for instance 57485f78-b97e-423c-b1f1-8648e963e5dd [ 858.581954] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5bdc3c4-b63c-46cc-b927-410c949c5346 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.599502] env[70013]: DEBUG oslo_vmware.api [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Task: {'id': task-4230904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206038} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.599502] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.599724] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 858.599837] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 858.600024] env[70013]: INFO nova.compute.manager [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Took 1.20 seconds to destroy the instance on the hypervisor. [ 858.600272] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 858.600458] env[70013]: DEBUG nova.compute.manager [-] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 858.600563] env[70013]: DEBUG nova.network.neutron [-] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 858.667837] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.668169] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.668365] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Deleting the datastore file [datastore1] 5c5fbdae-4916-4beb-ada0-57f36fb0f84c {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.668653] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df084a19-550d-4800-be08-c77febc83d20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.676704] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for the task: (returnval){ [ 858.676704] env[70013]: value = "task-4230906" [ 858.676704] env[70013]: _type = "Task" [ 858.676704] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.686991] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230906, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.763854] env[70013]: DEBUG nova.scheduler.client.report [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 858.781165] env[70013]: DEBUG nova.scheduler.client.report [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 858.781551] env[70013]: DEBUG nova.compute.provider_tree [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.798671] env[70013]: DEBUG nova.scheduler.client.report [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 858.818885] env[70013]: DEBUG nova.scheduler.client.report [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 858.884553] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Releasing lock "refresh_cache-477bf0d8-e988-43e9-a573-994a0d08df0c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 858.884833] env[70013]: DEBUG nova.compute.manager [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 858.884980] env[70013]: DEBUG nova.compute.manager [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 858.885166] env[70013]: DEBUG nova.network.neutron [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 858.909288] env[70013]: DEBUG nova.network.neutron [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.986651] env[70013]: DEBUG oslo_vmware.api [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230901, 'name': PowerOnVM_Task, 'duration_secs': 0.562062} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.990148] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.990596] env[70013]: INFO nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Took 8.04 seconds to spawn the instance on the hypervisor. [ 858.991011] env[70013]: DEBUG nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 859.002204] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a4a1a7-c85c-4ddb-8c9a-1acab627ebc8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.032086] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523ffddf-b2a2-f357-7eb9-31cd4fa607ea, 'name': SearchDatastore_Task, 'duration_secs': 0.012852} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.038197] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.038818] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.039188] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.039822] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 859.040237] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.041396] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23bfaef3-b32c-4e71-9bda-eb8acada3cc1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.057152] env[70013]: DEBUG nova.compute.manager [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 859.057152] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.063138] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1f11cd-a3ed-4cf6-8b60-afac34164929 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.069402] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.069886] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.072520] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f59fb454-19da-4c32-843f-99edd335afd4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.081529] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 859.082116] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.087582] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8daade69-379c-4e5c-8a23-050574b37a96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.090883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4f0cd3c6-21db-44ba-bdcc-aed4e17d82e7 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "57485f78-b97e-423c-b1f1-8648e963e5dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.253s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.092223] env[70013]: DEBUG nova.network.neutron [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Updated VIF entry in instance network info cache for port 3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.094144] env[70013]: DEBUG nova.network.neutron [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Updating instance_info_cache with network_info: [{"id": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "address": "fa:16:3e:1d:87:e4", "network": {"id": "a3991650-c4b9-4370-9a29-d0ad4de0877f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-440150710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47ea4715673458a82eedca76b444180", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0bd58d-e1", "ovs_interfaceid": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.100359] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 859.100359] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dff3df-a7ea-74c9-9a5c-1b68778e4ea4" [ 859.100359] env[70013]: _type = "Task" [ 859.100359] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.109309] env[70013]: DEBUG oslo_vmware.api [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 859.109309] env[70013]: value = "task-4230907" [ 859.109309] env[70013]: _type = "Task" [ 859.109309] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.124044] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dff3df-a7ea-74c9-9a5c-1b68778e4ea4, 'name': SearchDatastore_Task, 'duration_secs': 0.011378} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.130250] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11f32324-25be-45fe-9650-fec4fa37ba65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.140664] env[70013]: DEBUG oslo_vmware.api [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.148611] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 859.148611] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f5eef5-0b44-b790-9e6c-b5bd07f225dd" [ 859.148611] env[70013]: _type = "Task" [ 859.148611] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.165787] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f5eef5-0b44-b790-9e6c-b5bd07f225dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.195163] env[70013]: DEBUG oslo_vmware.api [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Task: {'id': task-4230906, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.333319} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.197443] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.197761] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 859.198129] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 859.293772] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 859.294234] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6c56a88-4111-4d8a-84b0-3cd7b87da978 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.308030] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f24fe54-4e35-4117-bfef-902c116a98cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.352402] env[70013]: ERROR nova.compute.manager [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Failed to detach volume 19da385b-04e0-4e04-9b4d-39b9370bec62 from /dev/sda: nova.exception.InstanceNotFound: Instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c could not be found. [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Traceback (most recent call last): [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 4185, in _do_rebuild_instance [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self.driver.rebuild(**kwargs) [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] raise NotImplementedError() [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] NotImplementedError [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] During handling of the above exception, another exception occurred: [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Traceback (most recent call last): [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3608, in _detach_root_volume [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self.driver.detach_volume(context, old_connection_info, [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] return self._volumeops.detach_volume(connection_info, instance) [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self._detach_volume_vmdk(connection_info, instance) [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] stable_ref.fetch_moref(session) [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] raise exception.InstanceNotFound(instance_id=self._uuid) [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] nova.exception.InstanceNotFound: Instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c could not be found. [ 859.352402] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.414548] env[70013]: DEBUG nova.network.neutron [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.434618] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87a3967e-130a-4764-a4fa-12239d9e9036 tempest-VolumesAssistedSnapshotsTest-347687562 tempest-VolumesAssistedSnapshotsTest-347687562-project-admin] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.317s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.525753] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35f41e2-5ded-4f09-8025-34359f2c4450 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.535342] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639331a9-5ea0-4164-9140-0436de6f554a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.547121] env[70013]: INFO nova.compute.manager [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Took 46.51 seconds to build instance. [ 859.584556] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9417864-eb3b-45da-a292-878b493b3512 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.594982] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aba991d-483b-4166-8f93-61eddf8c5ade {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.603538] env[70013]: DEBUG nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 859.607628] env[70013]: DEBUG oslo_concurrency.lockutils [req-7db26778-592b-4929-950b-340d806aa1b2 req-d5b3286e-6250-4972-b490-a7aef13c39c4 service nova] Releasing lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.620346] env[70013]: DEBUG nova.compute.provider_tree [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.624151] env[70013]: DEBUG nova.compute.utils [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Build of instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c aborted: Failed to rebuild volume backed instance. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 859.629891] env[70013]: ERROR nova.compute.manager [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c aborted: Failed to rebuild volume backed instance. [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Traceback (most recent call last): [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 4185, in _do_rebuild_instance [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self.driver.rebuild(**kwargs) [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] raise NotImplementedError() [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] NotImplementedError [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] During handling of the above exception, another exception occurred: [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Traceback (most recent call last): [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3643, in _rebuild_volume_backed_instance [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self._detach_root_volume(context, instance, root_bdm) [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3622, in _detach_root_volume [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] with excutils.save_and_reraise_exception(): [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self.force_reraise() [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] raise self.value [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3608, in _detach_root_volume [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self.driver.detach_volume(context, old_connection_info, [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] return self._volumeops.detach_volume(connection_info, instance) [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self._detach_volume_vmdk(connection_info, instance) [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] stable_ref.fetch_moref(session) [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] raise exception.InstanceNotFound(instance_id=self._uuid) [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] nova.exception.InstanceNotFound: Instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c could not be found. [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] During handling of the above exception, another exception occurred: [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Traceback (most recent call last): [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 11501, in _error_out_instance_on_exception [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] yield [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3911, in rebuild_instance [ 859.629891] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self._do_rebuild_instance_with_claim( [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3997, in _do_rebuild_instance_with_claim [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self._do_rebuild_instance( [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 4189, in _do_rebuild_instance [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self._rebuild_default_impl(**kwargs) [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3766, in _rebuild_default_impl [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] self._rebuild_volume_backed_instance( [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] File "/opt/stack/nova/nova/compute/manager.py", line 3658, in _rebuild_volume_backed_instance [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] raise exception.BuildAbortException( [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] nova.exception.BuildAbortException: Build of instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c aborted: Failed to rebuild volume backed instance. [ 859.631819] env[70013]: ERROR nova.compute.manager [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] [ 859.643153] env[70013]: DEBUG oslo_vmware.api [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230907, 'name': PowerOffVM_Task, 'duration_secs': 0.22316} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.643453] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.643653] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.643934] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef246ff2-51e3-4fd9-93d1-e34d03d3a919 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.659049] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f5eef5-0b44-b790-9e6c-b5bd07f225dd, 'name': SearchDatastore_Task, 'duration_secs': 0.012792} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.659342] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.659596] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268/6b458a7e-f786-41f6-b8ac-9f6f4ddcb268.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.659879] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-121d4614-e444-43f6-98ac-327a00a3b253 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.667856] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 859.667856] env[70013]: value = "task-4230909" [ 859.667856] env[70013]: _type = "Task" [ 859.667856] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.679885] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.701338] env[70013]: DEBUG nova.network.neutron [-] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.717392] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.717655] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.717870] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleting the datastore file [datastore2] f715cf45-81f1-4979-abfc-455b3e64f00e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.718167] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfc8196c-49e8-4052-962d-214298f750cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.727456] env[70013]: DEBUG oslo_vmware.api [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 859.727456] env[70013]: value = "task-4230910" [ 859.727456] env[70013]: _type = "Task" [ 859.727456] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.736689] env[70013]: DEBUG oslo_vmware.api [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230910, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.918060] env[70013]: INFO nova.compute.manager [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 477bf0d8-e988-43e9-a573-994a0d08df0c] Took 1.03 seconds to deallocate network for instance. [ 860.050130] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7351bf78-7314-4b32-b445-e10a6c3f3d6b tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.033s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.124136] env[70013]: DEBUG nova.scheduler.client.report [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 860.141401] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.179504] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230909, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.203745] env[70013]: INFO nova.compute.manager [-] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Took 1.60 seconds to deallocate network for instance. [ 860.241840] env[70013]: DEBUG oslo_vmware.api [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4230910, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19516} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.242291] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.242387] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.242501] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.242671] env[70013]: INFO nova.compute.manager [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Took 1.19 seconds to destroy the instance on the hypervisor. [ 860.242927] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 860.243165] env[70013]: DEBUG nova.compute.manager [-] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 860.243262] env[70013]: DEBUG nova.network.neutron [-] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.557637] env[70013]: DEBUG nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 860.632488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.904s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.633103] env[70013]: DEBUG nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 860.635571] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.647s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.683280] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230909, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517995} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.683280] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268/6b458a7e-f786-41f6-b8ac-9f6f4ddcb268.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.683280] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.683496] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-608d5681-ba53-44a6-8665-77aff408affe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.690465] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 860.690465] env[70013]: value = "task-4230912" [ 860.690465] env[70013]: _type = "Task" [ 860.690465] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.701371] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230912, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.712907] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.854025] env[70013]: DEBUG nova.compute.manager [req-ff078e40-1f77-4742-b6be-3964b78d177c req-0bc6e6f1-1512-4f94-9610-8d6bc1b469c4 service nova] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Received event network-vif-deleted-7634504a-a16c-44ea-9a82-e8f88bcc65ea {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 860.956630] env[70013]: INFO nova.scheduler.client.report [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Deleted allocations for instance 477bf0d8-e988-43e9-a573-994a0d08df0c [ 861.084513] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.084513] env[70013]: DEBUG nova.network.neutron [-] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.140670] env[70013]: DEBUG nova.compute.utils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 861.142754] env[70013]: DEBUG nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 861.142935] env[70013]: DEBUG nova.network.neutron [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 861.173504] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d422d9-5deb-4656-988b-ac4d6bf24dde {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.183156] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929da467-4c21-4c55-b4dc-36bae7415676 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.217590] env[70013]: DEBUG nova.policy [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de5b9d6af0874cac981d493abb184e04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65f99085562a43bd90cc165765a22dd7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 861.224546] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0ae625-ce20-4feb-ae2a-078dacd96129 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.236622] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230912, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068885} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.239585] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.240833] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45df38d-5203-47a7-b54a-32f3a4e39035 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.248115] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2649b5b6-24e8-4cfc-95a9-cffe664c0de9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.260882] env[70013]: DEBUG nova.compute.provider_tree [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.282924] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268/6b458a7e-f786-41f6-b8ac-9f6f4ddcb268.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.284381] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-661c6643-1f0f-46a4-973a-d9e9d144e0c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.308730] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 861.308730] env[70013]: value = "task-4230913" [ 861.308730] env[70013]: _type = "Task" [ 861.308730] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.317610] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.469039] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a995464d-78b0-4935-9f8c-d38236c91efc tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "477bf0d8-e988-43e9-a573-994a0d08df0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.768s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.588174] env[70013]: INFO nova.compute.manager [-] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Took 1.34 seconds to deallocate network for instance. [ 861.647942] env[70013]: DEBUG nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 861.651171] env[70013]: DEBUG oslo_concurrency.lockutils [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.806684] env[70013]: ERROR nova.scheduler.client.report [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [req-29be5cf7-4471-4d48-80f2-a31aa2dcb81d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-29be5cf7-4471-4d48-80f2-a31aa2dcb81d"}]}: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 861.807223] env[70013]: DEBUG oslo_concurrency.lockutils [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.172s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.808084] env[70013]: WARNING nova.compute.manager [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] [instance: f12b8575-5082-4be9-9bf5-f4279860d19d] Failed to revert task state for instance. Error: Failed to synchronize the placement service with resource provider information supplied by the compute host.: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 861.810782] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.104s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.811062] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.813777] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.871s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.815291] env[70013]: INFO nova.compute.claims [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server [None req-30f5425b-5e55-4aa1-b5e1-7cf4981c03a8 tempest-ImagesOneServerTestJSON-378500094 tempest-ImagesOneServerTestJSON-378500094-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server yield [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5b10b7b6-fd96-4bb9-b6c1-f8a8ba62e9ef"}]} [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 861.819053] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 861.822151] env[70013]: ERROR oslo_messaging.rpc.server [ 861.832044] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230913, 'name': ReconfigVM_Task, 'duration_secs': 0.292059} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.832389] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268/6b458a7e-f786-41f6-b8ac-9f6f4ddcb268.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.834056] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da226de7-6981-4864-87e6-9842e9ca2362 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.842300] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 861.842300] env[70013]: value = "task-4230914" [ 861.842300] env[70013]: _type = "Task" [ 861.842300] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.856174] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230914, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.869057] env[70013]: INFO nova.scheduler.client.report [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted allocations for instance 77c12460-9cfa-41c9-a210-238a470d9ccd [ 861.971541] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 861.975149] env[70013]: DEBUG nova.network.neutron [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Successfully created port: f3add360-018f-4a8b-9a58-1a47a83f2bc4 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.095956] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.208661] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquiring lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.208898] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.210983] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquiring lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.210983] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.210983] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.212189] env[70013]: INFO nova.compute.manager [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Terminating instance [ 862.357489] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230914, 'name': Rename_Task, 'duration_secs': 0.150888} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.358151] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.359123] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5260dc45-0ed1-47b7-b4fc-426d1650bda0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.367725] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 862.367725] env[70013]: value = "task-4230915" [ 862.367725] env[70013]: _type = "Task" [ 862.367725] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.382686] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.383785] env[70013]: DEBUG oslo_concurrency.lockutils [None req-075d339e-ce86-4c64-8c82-fe6a32bc23f3 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "77c12460-9cfa-41c9-a210-238a470d9ccd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.947s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.509258] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.627920] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.628220] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.628976] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.629242] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.629429] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.631721] env[70013]: INFO nova.compute.manager [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Terminating instance [ 862.658169] env[70013]: DEBUG nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 862.702464] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 862.702842] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.702962] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 862.703457] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.703797] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 862.703900] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 862.704283] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 862.704283] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 862.704459] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 862.704607] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 862.704930] env[70013]: DEBUG nova.virt.hardware [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 862.705929] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0daa50d-28da-49d9-b657-fd81f07b2045 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.715716] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828713b1-6f44-4f9b-9509-104ac3abbb3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.722293] env[70013]: DEBUG nova.compute.manager [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 862.722891] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f597321-71c7-4c7f-9ae4-102114e46ecf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.742490] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409f0148-e8cc-446a-a630-f026e2822813 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.780162] env[70013]: WARNING nova.virt.vmwareapi.driver [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c could not be found. [ 862.780423] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 862.784320] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-662f05bb-0829-410e-a365-7d3ca6717809 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.791422] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "30fcb8c7-0947-426e-9aa5-081446e31c28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.791645] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.798932] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264210d5-93e5-49a1-bc6c-9ffbc0f0e8a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.835710] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c could not be found. [ 862.835907] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 862.836112] env[70013]: INFO nova.compute.manager [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Took 0.11 seconds to destroy the instance on the hypervisor. [ 862.836365] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 862.837881] env[70013]: DEBUG nova.compute.manager [-] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 862.838888] env[70013]: DEBUG nova.network.neutron [-] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 862.859728] env[70013]: DEBUG nova.scheduler.client.report [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 862.879750] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230915, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.881969] env[70013]: DEBUG nova.scheduler.client.report [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 862.881969] env[70013]: DEBUG nova.compute.provider_tree [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 862.902729] env[70013]: DEBUG nova.scheduler.client.report [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 862.940537] env[70013]: DEBUG nova.scheduler.client.report [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 863.010556] env[70013]: DEBUG nova.compute.manager [req-6a140d1f-2c02-4e01-ba55-9d07c1fd0427 req-e973c8ec-e37b-4623-b2ca-e6ea6c8d1616 service nova] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Received event network-vif-deleted-c0c18e7a-0f35-4d2e-b12d-0002001b6793 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 863.136284] env[70013]: DEBUG nova.compute.manager [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 863.136561] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.137860] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6f066c-e933-467a-b8b6-5aed5b90e051 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.146492] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.151027] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6df8bb3-32e5-46dc-a0a4-06cb519bb7e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.160035] env[70013]: DEBUG oslo_vmware.api [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 863.160035] env[70013]: value = "task-4230916" [ 863.160035] env[70013]: _type = "Task" [ 863.160035] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.169441] env[70013]: DEBUG oslo_vmware.api [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230916, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.385110] env[70013]: DEBUG oslo_vmware.api [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230915, 'name': PowerOnVM_Task, 'duration_secs': 0.6041} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.385449] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.386294] env[70013]: INFO nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Took 8.42 seconds to spawn the instance on the hypervisor. [ 863.386294] env[70013]: DEBUG nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 863.389551] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17246eb-8c94-42b9-86c4-bec31d4f4fe9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.439637] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8727079f-e846-4085-9aec-f2555e541f75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.450233] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24831035-3128-4e64-b110-45210fbf43dc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.455323] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "be9a565b-c74c-442e-8087-cc4579b1c8e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.455652] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.461617] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "be9a565b-c74c-442e-8087-cc4579b1c8e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.461617] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.461617] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.466700] env[70013]: INFO nova.compute.manager [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Terminating instance [ 863.494882] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1da5a6e-c49a-41b6-af8c-3896e2884776 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.506273] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a3cf46-41d1-432a-a35f-a3d8bfab4fce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.524996] env[70013]: DEBUG nova.compute.provider_tree [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 863.667755] env[70013]: DEBUG oslo_vmware.api [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230916, 'name': PowerOffVM_Task, 'duration_secs': 0.390277} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.671022] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.671022] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 863.671022] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67453a01-b61c-4870-9994-a59ea8628ffb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.699041] env[70013]: DEBUG nova.network.neutron [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Successfully updated port: f3add360-018f-4a8b-9a58-1a47a83f2bc4 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.754321] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 863.754640] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 863.754922] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Deleting the datastore file [datastore1] b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.755459] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7879f7e7-ec06-4925-87ee-006a070ab959 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.766085] env[70013]: DEBUG oslo_vmware.api [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for the task: (returnval){ [ 863.766085] env[70013]: value = "task-4230918" [ 863.766085] env[70013]: _type = "Task" [ 863.766085] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.775134] env[70013]: DEBUG oslo_vmware.api [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230918, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.908616] env[70013]: INFO nova.compute.manager [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Took 48.77 seconds to build instance. [ 864.000709] env[70013]: DEBUG nova.compute.manager [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 864.000709] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.003773] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04883ba-3b67-43d4-ab75-da372e5a408b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.015107] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.015107] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa15aa0f-6b46-4983-81df-27a947cd276d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.021281] env[70013]: DEBUG oslo_vmware.api [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 864.021281] env[70013]: value = "task-4230919" [ 864.021281] env[70013]: _type = "Task" [ 864.021281] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.034043] env[70013]: DEBUG oslo_vmware.api [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.063248] env[70013]: ERROR nova.scheduler.client.report [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [req-3f6ab06d-81ed-40c2-bbbb-b9a04ea9ec17] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3f6ab06d-81ed-40c2-bbbb-b9a04ea9ec17"}]} [ 864.063248] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.063433] env[70013]: ERROR nova.compute.manager [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Traceback (most recent call last): [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] yield [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] self.set_inventory_for_provider( [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3f6ab06d-81ed-40c2-bbbb-b9a04ea9ec17"}]} [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] During handling of the above exception, another exception occurred: [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Traceback (most recent call last): [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] with self.rt.instance_claim(context, instance, node, allocs, [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] return f(*args, **kwargs) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] self._update(elevated, cn) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] self._update_to_placement(context, compute_node, startup) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] return attempt.get(self._wrap_exception) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] raise value [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] self.reportclient.update_from_provider_tree( [ 864.063433] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 864.064430] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] with catch_all(pd.uuid): [ 864.064430] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 864.064430] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] self.gen.throw(typ, value, traceback) [ 864.064430] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 864.064430] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] raise exception.ResourceProviderSyncFailed() [ 864.064430] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 864.064430] env[70013]: ERROR nova.compute.manager [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] [ 864.064430] env[70013]: DEBUG nova.compute.utils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 864.066091] env[70013]: DEBUG nova.compute.manager [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Build of instance c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 864.068271] env[70013]: DEBUG nova.compute.manager [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 864.068271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "refresh_cache-c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.068271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquired lock "refresh_cache-c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.068271] env[70013]: DEBUG nova.network.neutron [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 864.069855] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.261s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 864.071505] env[70013]: INFO nova.compute.claims [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.210133] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "refresh_cache-0b0b2769-74ba-4ba2-9e8b-148cc2976321" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.210133] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquired lock "refresh_cache-0b0b2769-74ba-4ba2-9e8b-148cc2976321" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.210133] env[70013]: DEBUG nova.network.neutron [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 864.276863] env[70013]: DEBUG oslo_vmware.api [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Task: {'id': task-4230918, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181223} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.277194] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.277396] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.277608] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.277736] env[70013]: INFO nova.compute.manager [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Took 1.14 seconds to destroy the instance on the hypervisor. [ 864.278245] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 864.278487] env[70013]: DEBUG nova.compute.manager [-] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 864.278701] env[70013]: DEBUG nova.network.neutron [-] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.315149] env[70013]: DEBUG nova.compute.manager [req-6f6cd783-54cd-47dd-b2a5-ed28170a7c4e req-675182ca-aa24-45d4-8aa3-4f3430a83793 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Received event network-vif-deleted-2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 864.316029] env[70013]: INFO nova.compute.manager [req-6f6cd783-54cd-47dd-b2a5-ed28170a7c4e req-675182ca-aa24-45d4-8aa3-4f3430a83793 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Neutron deleted interface 2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8; detaching it from the instance and deleting it from the info cache [ 864.316029] env[70013]: DEBUG nova.network.neutron [req-6f6cd783-54cd-47dd-b2a5-ed28170a7c4e req-675182ca-aa24-45d4-8aa3-4f3430a83793 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.411082] env[70013]: DEBUG oslo_concurrency.lockutils [None req-edb1ed35-5bf2-44f3-9128-c98e6322c7af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.254s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.532545] env[70013]: DEBUG oslo_vmware.api [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230919, 'name': PowerOffVM_Task, 'duration_secs': 0.211876} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.532810] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.532983] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.533356] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8f8ef2f-bbab-4ed0-aea5-5717b5501cc8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.584385] env[70013]: DEBUG nova.network.neutron [-] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.600809] env[70013]: DEBUG nova.network.neutron [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.607338] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.609652] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.611720] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Deleting the datastore file [datastore2] be9a565b-c74c-442e-8087-cc4579b1c8e7 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.611720] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5922e77a-0d27-4c66-ae61-e981babca7e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.619441] env[70013]: DEBUG oslo_vmware.api [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 864.619441] env[70013]: value = "task-4230921" [ 864.619441] env[70013]: _type = "Task" [ 864.619441] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.629870] env[70013]: DEBUG oslo_vmware.api [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.696622] env[70013]: DEBUG nova.network.neutron [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.779654] env[70013]: DEBUG nova.network.neutron [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.818109] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d7798aa-ffeb-4e42-8e4a-60cfcdd51dcc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.832789] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781ffdce-bc5b-4ae1-9a88-9789b4fe466a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.867224] env[70013]: DEBUG nova.compute.manager [req-6f6cd783-54cd-47dd-b2a5-ed28170a7c4e req-675182ca-aa24-45d4-8aa3-4f3430a83793 service nova] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Detach interface failed, port_id=2cd50eea-d7e9-4cb1-9ca6-65573c52b7e8, reason: Instance 5c5fbdae-4916-4beb-ada0-57f36fb0f84c could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 864.915593] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 865.006662] env[70013]: DEBUG nova.network.neutron [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Updating instance_info_cache with network_info: [{"id": "f3add360-018f-4a8b-9a58-1a47a83f2bc4", "address": "fa:16:3e:02:3d:8d", "network": {"id": "c36d49c4-78f7-49a3-94b8-a5d66daff1d8", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-807374359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65f99085562a43bd90cc165765a22dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3add360-01", "ovs_interfaceid": "f3add360-018f-4a8b-9a58-1a47a83f2bc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.093371] env[70013]: INFO nova.compute.manager [-] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Took 2.26 seconds to deallocate network for instance. [ 865.112539] env[70013]: DEBUG nova.scheduler.client.report [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 865.132687] env[70013]: DEBUG oslo_vmware.api [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4230921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.49612} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.133727] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.133974] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.134204] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.134428] env[70013]: INFO nova.compute.manager [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 865.134719] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 865.135877] env[70013]: DEBUG nova.scheduler.client.report [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 865.136363] env[70013]: DEBUG nova.compute.provider_tree [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 865.139888] env[70013]: DEBUG nova.compute.manager [-] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 865.140074] env[70013]: DEBUG nova.network.neutron [-] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.153276] env[70013]: DEBUG nova.scheduler.client.report [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 865.174391] env[70013]: DEBUG nova.scheduler.client.report [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 865.201518] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Releasing lock "refresh_cache-c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.201809] env[70013]: DEBUG nova.compute.manager [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 865.201809] env[70013]: DEBUG nova.compute.manager [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 865.202074] env[70013]: DEBUG nova.network.neutron [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.222123] env[70013]: DEBUG oslo_concurrency.lockutils [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.222293] env[70013]: DEBUG oslo_concurrency.lockutils [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.222514] env[70013]: INFO nova.compute.manager [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Rebooting instance [ 865.249995] env[70013]: DEBUG nova.network.neutron [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.450429] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.512857] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Releasing lock "refresh_cache-0b0b2769-74ba-4ba2-9e8b-148cc2976321" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.513144] env[70013]: DEBUG nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Instance network_info: |[{"id": "f3add360-018f-4a8b-9a58-1a47a83f2bc4", "address": "fa:16:3e:02:3d:8d", "network": {"id": "c36d49c4-78f7-49a3-94b8-a5d66daff1d8", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-807374359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65f99085562a43bd90cc165765a22dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3add360-01", "ovs_interfaceid": "f3add360-018f-4a8b-9a58-1a47a83f2bc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 865.517321] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:3d:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3add360-018f-4a8b-9a58-1a47a83f2bc4', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.523882] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Creating folder: Project (65f99085562a43bd90cc165765a22dd7). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.527190] env[70013]: DEBUG nova.compute.manager [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Received event network-vif-plugged-f3add360-018f-4a8b-9a58-1a47a83f2bc4 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 865.527447] env[70013]: DEBUG oslo_concurrency.lockutils [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] Acquiring lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.527728] env[70013]: DEBUG oslo_concurrency.lockutils [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.527893] env[70013]: DEBUG oslo_concurrency.lockutils [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.528078] env[70013]: DEBUG nova.compute.manager [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] No waiting events found dispatching network-vif-plugged-f3add360-018f-4a8b-9a58-1a47a83f2bc4 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 865.528243] env[70013]: WARNING nova.compute.manager [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Received unexpected event network-vif-plugged-f3add360-018f-4a8b-9a58-1a47a83f2bc4 for instance with vm_state building and task_state spawning. [ 865.528848] env[70013]: DEBUG nova.compute.manager [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Received event network-changed-f3add360-018f-4a8b-9a58-1a47a83f2bc4 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 865.528848] env[70013]: DEBUG nova.compute.manager [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Refreshing instance network info cache due to event network-changed-f3add360-018f-4a8b-9a58-1a47a83f2bc4. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 865.528848] env[70013]: DEBUG oslo_concurrency.lockutils [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] Acquiring lock "refresh_cache-0b0b2769-74ba-4ba2-9e8b-148cc2976321" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.528848] env[70013]: DEBUG oslo_concurrency.lockutils [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] Acquired lock "refresh_cache-0b0b2769-74ba-4ba2-9e8b-148cc2976321" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.528982] env[70013]: DEBUG nova.network.neutron [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Refreshing network info cache for port f3add360-018f-4a8b-9a58-1a47a83f2bc4 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.532694] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9193b15f-a0c5-4efd-83dd-d1a2f3b62542 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.553039] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Created folder: Project (65f99085562a43bd90cc165765a22dd7) in parent group-v836999. [ 865.553271] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Creating folder: Instances. Parent ref: group-v837119. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.554127] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95723e44-1615-49b8-a2c5-85dd38b560d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.571820] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Created folder: Instances in parent group-v837119. [ 865.571820] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 865.575235] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.577069] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7cff599-4fc1-44cd-b3d8-f06e667bbca4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.601922] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.601922] env[70013]: value = "task-4230924" [ 865.601922] env[70013]: _type = "Task" [ 865.601922] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.615487] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230924, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.680020] env[70013]: INFO nova.compute.manager [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Took 0.58 seconds to detach 1 volumes for instance. [ 865.681312] env[70013]: DEBUG nova.compute.manager [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Deleting volume: 19da385b-04e0-4e04-9b4d-39b9370bec62 {{(pid=70013) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 865.754053] env[70013]: DEBUG nova.network.neutron [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.757964] env[70013]: DEBUG oslo_concurrency.lockutils [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.758259] env[70013]: DEBUG oslo_concurrency.lockutils [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquired lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.758438] env[70013]: DEBUG nova.network.neutron [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.810314] env[70013]: DEBUG nova.network.neutron [-] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.850838] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd832360-6b59-4572-be8c-3b7fc2ec8414 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.859981] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f156d299-752e-41ef-afcc-988ebec88b8d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.900640] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7586ce97-9b63-4fc1-9f18-6b52a88b7d27 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.910481] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c26e36a-411f-4c5b-95f6-40bba5d2e093 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.927982] env[70013]: DEBUG nova.compute.provider_tree [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.118367] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230924, 'name': CreateVM_Task, 'duration_secs': 0.37255} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.118650] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.119821] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.119876] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.120462] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 866.123948] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a6d7e9f-333e-4b3f-8c68-3047089c7154 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.127511] env[70013]: DEBUG nova.network.neutron [-] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.136717] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 866.136717] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b6fca8-c606-d8b8-24c2-87b5c685cdd0" [ 866.136717] env[70013]: _type = "Task" [ 866.136717] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.148694] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b6fca8-c606-d8b8-24c2-87b5c685cdd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.240827] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.261891] env[70013]: INFO nova.compute.manager [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3] Took 1.06 seconds to deallocate network for instance. [ 866.313822] env[70013]: INFO nova.compute.manager [-] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Took 2.03 seconds to deallocate network for instance. [ 866.365845] env[70013]: DEBUG nova.compute.manager [req-c13490b3-34ae-41c6-aff2-80ae957ad350 req-c74e7b89-1898-4324-a2e7-c5a6c8f3e28c service nova] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Received event network-vif-deleted-47970e5c-2dab-40b7-abcc-f2bc90bcfad7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 866.365845] env[70013]: DEBUG nova.network.neutron [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Updated VIF entry in instance network info cache for port f3add360-018f-4a8b-9a58-1a47a83f2bc4. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 866.367644] env[70013]: DEBUG nova.network.neutron [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Updating instance_info_cache with network_info: [{"id": "f3add360-018f-4a8b-9a58-1a47a83f2bc4", "address": "fa:16:3e:02:3d:8d", "network": {"id": "c36d49c4-78f7-49a3-94b8-a5d66daff1d8", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-807374359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65f99085562a43bd90cc165765a22dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3add360-01", "ovs_interfaceid": "f3add360-018f-4a8b-9a58-1a47a83f2bc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.432190] env[70013]: DEBUG nova.scheduler.client.report [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 866.632453] env[70013]: INFO nova.compute.manager [-] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Took 1.49 seconds to deallocate network for instance. [ 866.649967] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b6fca8-c606-d8b8-24c2-87b5c685cdd0, 'name': SearchDatastore_Task, 'duration_secs': 0.017062} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.650302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.650853] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.651012] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.651176] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.651356] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.651635] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d04b8cd-a881-40c1-b93c-196337accc45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.662729] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.662967] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.663696] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-369a7ef1-36c2-475a-be54-aff4eaebed41 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.671261] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 866.671261] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3b054-99db-7091-a191-0fe1f1521843" [ 866.671261] env[70013]: _type = "Task" [ 866.671261] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.680823] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3b054-99db-7091-a191-0fe1f1521843, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.798347] env[70013]: DEBUG nova.network.neutron [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Updating instance_info_cache with network_info: [{"id": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "address": "fa:16:3e:1d:87:e4", "network": {"id": "a3991650-c4b9-4370-9a29-d0ad4de0877f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-440150710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47ea4715673458a82eedca76b444180", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0bd58d-e1", "ovs_interfaceid": "3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.822661] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.870133] env[70013]: DEBUG oslo_concurrency.lockutils [req-c0c4b86c-48e3-4fdb-87b3-6b56955ee929 req-40bd5655-1a50-4bca-923a-fe32b89a6450 service nova] Releasing lock "refresh_cache-0b0b2769-74ba-4ba2-9e8b-148cc2976321" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.938961] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.869s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 866.939088] env[70013]: DEBUG nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 866.944633] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.842s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.946910] env[70013]: INFO nova.compute.claims [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.143828] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.186612] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3b054-99db-7091-a191-0fe1f1521843, 'name': SearchDatastore_Task, 'duration_secs': 0.032874} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.187668] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95cc0042-5e52-4e44-8ad5-407fa832402f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.194834] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 867.194834] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e4a964-7509-fea5-27e9-d0841651c02b" [ 867.194834] env[70013]: _type = "Task" [ 867.194834] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.205582] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e4a964-7509-fea5-27e9-d0841651c02b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.299970] env[70013]: INFO nova.scheduler.client.report [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Deleted allocations for instance c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3 [ 867.305466] env[70013]: DEBUG oslo_concurrency.lockutils [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Releasing lock "refresh_cache-6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.445579] env[70013]: DEBUG nova.compute.utils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 867.447899] env[70013]: DEBUG nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 867.447899] env[70013]: DEBUG nova.network.neutron [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 867.551705] env[70013]: DEBUG nova.policy [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '99104b73f6174c05a5272f023612bb49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f85175ef2b004d5ea372f2dfe69549ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 867.556793] env[70013]: DEBUG nova.compute.manager [req-258f2034-fa64-498b-9a19-c00e5277a36d req-38a8a4d4-68b5-4a34-a6dd-56c246afa809 service nova] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Received event network-vif-deleted-84120f9b-ce4d-4d5a-9090-faca58cc4f6a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 867.707179] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e4a964-7509-fea5-27e9-d0841651c02b, 'name': SearchDatastore_Task, 'duration_secs': 0.017468} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.707542] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.707752] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0b0b2769-74ba-4ba2-9e8b-148cc2976321/0b0b2769-74ba-4ba2-9e8b-148cc2976321.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.708103] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96738c6b-0ff7-4b80-8342-d2f79f88639e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.715905] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 867.715905] env[70013]: value = "task-4230926" [ 867.715905] env[70013]: _type = "Task" [ 867.715905] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.725542] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230926, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.814107] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2f3e43bb-f104-4a64-8399-e75343004d07 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "c1aacf70-a6b5-4528-bcaa-b8f1c955a1c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.969s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.814107] env[70013]: DEBUG nova.compute.manager [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 867.814107] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fe28c7-1776-420a-a582-9354a7c46859 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.950847] env[70013]: DEBUG nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 868.235536] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230926, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.269032] env[70013]: DEBUG nova.network.neutron [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Successfully created port: 95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.319324] env[70013]: DEBUG nova.compute.manager [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 868.617784] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5004aefa-2c48-4cf3-baf7-3079268e4803 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.632469] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75962164-5aa8-4c92-812e-a5361f22fad1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.673721] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3e3e45-9ce6-4b51-b38e-63637408faab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.682094] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0ac8e8-e1af-4607-8163-83288a792ee9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.699233] env[70013]: DEBUG nova.compute.provider_tree [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.727934] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230926, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516118} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.728047] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0b0b2769-74ba-4ba2-9e8b-148cc2976321/0b0b2769-74ba-4ba2-9e8b-148cc2976321.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.728268] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.728531] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f506263-df46-47e4-8226-349ce62cb83b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.737298] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 868.737298] env[70013]: value = "task-4230927" [ 868.737298] env[70013]: _type = "Task" [ 868.737298] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.747585] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230927, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.834655] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11c8b04-f0c9-4185-a870-b25c322b232b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.843103] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Doing hard reboot of VM {{(pid=70013) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 868.843396] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-b7a39376-5f4a-4ee4-be31-a90c7b061644 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.846634] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.853211] env[70013]: DEBUG oslo_vmware.api [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 868.853211] env[70013]: value = "task-4230928" [ 868.853211] env[70013]: _type = "Task" [ 868.853211] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.864462] env[70013]: DEBUG oslo_vmware.api [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230928, 'name': ResetVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.968956] env[70013]: DEBUG nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 869.007036] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 869.007036] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.007036] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 869.007036] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.007036] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 869.007036] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 869.007443] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 869.007772] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 869.008100] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 869.010132] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 869.010132] env[70013]: DEBUG nova.virt.hardware [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 869.010132] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876204b5-9b8e-45b8-8a3f-2c8f1dc299b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.018803] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1a95ec-bfbd-4edf-a98c-c8ed0749fa7a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.202941] env[70013]: DEBUG nova.scheduler.client.report [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.251619] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230927, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.196478} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.252019] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.253184] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bf521c-8081-4404-bb62-b36a94c0ab9a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.291089] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 0b0b2769-74ba-4ba2-9e8b-148cc2976321/0b0b2769-74ba-4ba2-9e8b-148cc2976321.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.291498] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bad1f24e-1c43-478b-ba06-ae6e26889271 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.322521] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 869.322521] env[70013]: value = "task-4230929" [ 869.322521] env[70013]: _type = "Task" [ 869.322521] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.335345] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230929, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.366612] env[70013]: DEBUG oslo_vmware.api [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230928, 'name': ResetVM_Task, 'duration_secs': 0.131913} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.367075] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Did hard reboot of VM {{(pid=70013) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 869.367437] env[70013]: DEBUG nova.compute.manager [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 869.368645] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8753e377-8a6e-43b6-8d9c-9027260d6d76 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.622921] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "3b29c99a-0d55-40b1-a155-199b1ced2146" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.622921] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.622921] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.622921] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.622921] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.624435] env[70013]: INFO nova.compute.manager [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Terminating instance [ 869.713023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.767s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.713023] env[70013]: DEBUG nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 869.713996] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.147s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.714381] env[70013]: DEBUG nova.objects.instance [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lazy-loading 'resources' on Instance uuid 352790fe-057f-45ee-aff0-549b456fd181 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.833640] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230929, 'name': ReconfigVM_Task, 'duration_secs': 0.376628} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.834353] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 0b0b2769-74ba-4ba2-9e8b-148cc2976321/0b0b2769-74ba-4ba2-9e8b-148cc2976321.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.835103] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-252ac108-113b-47d1-9223-ffa379e0604a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.846271] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 869.846271] env[70013]: value = "task-4230930" [ 869.846271] env[70013]: _type = "Task" [ 869.846271] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.856999] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230930, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.885032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-278cf679-8ded-4446-bcfa-f02bb041a8af tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.663s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.000543] env[70013]: DEBUG nova.compute.manager [req-9da777b6-2b4f-4609-8b3a-6e62ad6e207e req-7b9da76c-5e0b-416d-ab50-ab36c6141e70 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Received event network-vif-plugged-95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 870.000945] env[70013]: DEBUG oslo_concurrency.lockutils [req-9da777b6-2b4f-4609-8b3a-6e62ad6e207e req-7b9da76c-5e0b-416d-ab50-ab36c6141e70 service nova] Acquiring lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.001207] env[70013]: DEBUG oslo_concurrency.lockutils [req-9da777b6-2b4f-4609-8b3a-6e62ad6e207e req-7b9da76c-5e0b-416d-ab50-ab36c6141e70 service nova] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 870.001502] env[70013]: DEBUG oslo_concurrency.lockutils [req-9da777b6-2b4f-4609-8b3a-6e62ad6e207e req-7b9da76c-5e0b-416d-ab50-ab36c6141e70 service nova] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.001775] env[70013]: DEBUG nova.compute.manager [req-9da777b6-2b4f-4609-8b3a-6e62ad6e207e req-7b9da76c-5e0b-416d-ab50-ab36c6141e70 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] No waiting events found dispatching network-vif-plugged-95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 870.001998] env[70013]: WARNING nova.compute.manager [req-9da777b6-2b4f-4609-8b3a-6e62ad6e207e req-7b9da76c-5e0b-416d-ab50-ab36c6141e70 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Received unexpected event network-vif-plugged-95450791-e5ab-4650-be74-3d6299da9718 for instance with vm_state building and task_state spawning. [ 870.129841] env[70013]: DEBUG nova.compute.manager [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 870.129841] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.130964] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16eb4a1-c2e3-48de-b5b7-73dc6612f784 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.141318] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.141380] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f5a2110-9515-4f90-b9a4-9340780651fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.150019] env[70013]: DEBUG oslo_vmware.api [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 870.150019] env[70013]: value = "task-4230931" [ 870.150019] env[70013]: _type = "Task" [ 870.150019] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.159496] env[70013]: DEBUG oslo_vmware.api [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.215793] env[70013]: DEBUG nova.compute.utils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 870.217574] env[70013]: DEBUG nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 870.217818] env[70013]: DEBUG nova.network.neutron [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 870.299392] env[70013]: DEBUG nova.policy [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a8d904839254e55bd56cd8b87b93711', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '208376e0d950450cb2ce841e2d049cd2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 870.358585] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230930, 'name': Rename_Task, 'duration_secs': 0.160464} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.358989] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.359334] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1f0bc0c-8740-4d2e-a40d-e45a041dbb99 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.370764] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 870.370764] env[70013]: value = "task-4230932" [ 870.370764] env[70013]: _type = "Task" [ 870.370764] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.381067] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230932, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.435198] env[70013]: DEBUG nova.network.neutron [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Successfully updated port: 95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.662795] env[70013]: DEBUG oslo_vmware.api [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230931, 'name': PowerOffVM_Task, 'duration_secs': 0.278224} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.665370] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.665742] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.668599] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce544905-13f9-425f-8b10-26d16d4d267a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.726409] env[70013]: DEBUG nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 870.751433] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.751433] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.752859] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Deleting the datastore file [datastore1] 3b29c99a-0d55-40b1-a155-199b1ced2146 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.756298] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e108bd59-6c9b-4ac8-a5c3-a51ff93689f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.764926] env[70013]: DEBUG oslo_vmware.api [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for the task: (returnval){ [ 870.764926] env[70013]: value = "task-4230934" [ 870.764926] env[70013]: _type = "Task" [ 870.764926] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.777902] env[70013]: DEBUG oslo_vmware.api [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230934, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.817042] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cabb530-2f29-484c-958d-3a4a63b5de8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.829685] env[70013]: DEBUG nova.network.neutron [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Successfully created port: 4a9242b0-4afa-443e-9ccd-dc35313fc59a {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.834551] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df36db1f-a5e7-492a-b793-8db6e606f93c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.887254] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d4df28-10af-4d8f-8f4e-0f1710d8abe6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.897153] env[70013]: DEBUG oslo_vmware.api [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230932, 'name': PowerOnVM_Task, 'duration_secs': 0.513606} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.899804] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.903329] env[70013]: INFO nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Took 8.24 seconds to spawn the instance on the hypervisor. [ 870.904663] env[70013]: DEBUG nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 870.904663] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d424b501-bbef-456d-8b0b-8a48437a2b46 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.908764] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfe2b85-0770-47ec-bdf8-f49c085e5888 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.932050] env[70013]: DEBUG nova.compute.provider_tree [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.940659] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.940659] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 870.940659] env[70013]: DEBUG nova.network.neutron [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.053608] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.053764] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.053988] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.054185] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.054356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.060468] env[70013]: INFO nova.compute.manager [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Terminating instance [ 871.277157] env[70013]: DEBUG oslo_vmware.api [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Task: {'id': task-4230934, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.327988} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.277471] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.277620] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.278105] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.278105] env[70013]: INFO nova.compute.manager [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Took 1.15 seconds to destroy the instance on the hypervisor. [ 871.278236] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 871.278428] env[70013]: DEBUG nova.compute.manager [-] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 871.278518] env[70013]: DEBUG nova.network.neutron [-] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 871.442147] env[70013]: DEBUG nova.scheduler.client.report [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.456241] env[70013]: INFO nova.compute.manager [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Took 49.70 seconds to build instance. [ 871.513774] env[70013]: DEBUG nova.network.neutron [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.567027] env[70013]: DEBUG nova.compute.manager [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 871.567027] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.567027] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d380515c-4918-424a-a437-67c9e55ed9fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.580070] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.583380] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-267162a9-0153-4ff3-9371-67b31acd1842 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.591724] env[70013]: DEBUG oslo_vmware.api [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 871.591724] env[70013]: value = "task-4230935" [ 871.591724] env[70013]: _type = "Task" [ 871.591724] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.604494] env[70013]: DEBUG oslo_vmware.api [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230935, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.722342] env[70013]: DEBUG nova.network.neutron [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updating instance_info_cache with network_info: [{"id": "95450791-e5ab-4650-be74-3d6299da9718", "address": "fa:16:3e:39:92:a5", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95450791-e5", "ovs_interfaceid": "95450791-e5ab-4650-be74-3d6299da9718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.741683] env[70013]: DEBUG nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 871.774689] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 871.775011] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.775185] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 871.775371] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.775518] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 871.775666] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 871.776641] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 871.776902] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 871.777295] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 871.777595] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 871.777857] env[70013]: DEBUG nova.virt.hardware [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 871.779786] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9a018c-3cf6-4135-a168-bf6cdffa8701 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.795956] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be6cc2a-d0d5-4bb6-8d52-5af4b8c02b43 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.961034] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.245s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.963529] env[70013]: DEBUG oslo_concurrency.lockutils [None req-def78012-ddff-452e-a904-9121671cbcb1 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.344s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.964671] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.380s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.965127] env[70013]: DEBUG nova.objects.instance [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lazy-loading 'resources' on Instance uuid 02ef7875-7ec9-4409-aaa5-71ed669f3780 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.994766] env[70013]: INFO nova.scheduler.client.report [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Deleted allocations for instance 352790fe-057f-45ee-aff0-549b456fd181 [ 872.103198] env[70013]: DEBUG oslo_vmware.api [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230935, 'name': PowerOffVM_Task, 'duration_secs': 0.216737} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.103519] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.103703] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.103962] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ace2395a-b44f-4871-9ff5-1cdaf0e3938a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.110072] env[70013]: DEBUG nova.compute.manager [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Received event network-changed-95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 872.110072] env[70013]: DEBUG nova.compute.manager [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Refreshing instance network info cache due to event network-changed-95450791-e5ab-4650-be74-3d6299da9718. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 872.110365] env[70013]: DEBUG oslo_concurrency.lockutils [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] Acquiring lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.170231] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.170448] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.170704] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Deleting the datastore file [datastore1] 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.171113] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b30f8215-566c-4974-8c88-b6fb972e90a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.178716] env[70013]: DEBUG oslo_vmware.api [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for the task: (returnval){ [ 872.178716] env[70013]: value = "task-4230937" [ 872.178716] env[70013]: _type = "Task" [ 872.178716] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.190202] env[70013]: DEBUG nova.network.neutron [-] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.191473] env[70013]: DEBUG oslo_vmware.api [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230937, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.225379] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Releasing lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.225737] env[70013]: DEBUG nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Instance network_info: |[{"id": "95450791-e5ab-4650-be74-3d6299da9718", "address": "fa:16:3e:39:92:a5", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95450791-e5", "ovs_interfaceid": "95450791-e5ab-4650-be74-3d6299da9718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 872.226342] env[70013]: DEBUG oslo_concurrency.lockutils [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] Acquired lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 872.226342] env[70013]: DEBUG nova.network.neutron [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Refreshing network info cache for port 95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.227559] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:92:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95450791-e5ab-4650-be74-3d6299da9718', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.236015] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 872.236501] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.236733] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45048f99-c87d-4015-91e0-b7c778a8c254 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.258238] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.258238] env[70013]: value = "task-4230938" [ 872.258238] env[70013]: _type = "Task" [ 872.258238] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.270239] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230938, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.476991] env[70013]: DEBUG nova.compute.manager [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 872.507956] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4ea97f6-4f61-463f-ab57-55e5b949fae8 tempest-ListImageFiltersTestJSON-570141788 tempest-ListImageFiltersTestJSON-570141788-project-member] Lock "352790fe-057f-45ee-aff0-549b456fd181" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.768s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.605815] env[70013]: DEBUG nova.compute.manager [req-0debad7a-84cd-4b93-a1e4-a6e98502eb0f req-7ee91a4f-59ac-4ba8-a204-49a61c273202 service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Received event network-vif-plugged-4a9242b0-4afa-443e-9ccd-dc35313fc59a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 872.606455] env[70013]: DEBUG oslo_concurrency.lockutils [req-0debad7a-84cd-4b93-a1e4-a6e98502eb0f req-7ee91a4f-59ac-4ba8-a204-49a61c273202 service nova] Acquiring lock "789fffd9-6725-4bf6-9144-dd603b0a521f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.606722] env[70013]: DEBUG oslo_concurrency.lockutils [req-0debad7a-84cd-4b93-a1e4-a6e98502eb0f req-7ee91a4f-59ac-4ba8-a204-49a61c273202 service nova] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 872.606896] env[70013]: DEBUG oslo_concurrency.lockutils [req-0debad7a-84cd-4b93-a1e4-a6e98502eb0f req-7ee91a4f-59ac-4ba8-a204-49a61c273202 service nova] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.607326] env[70013]: DEBUG nova.compute.manager [req-0debad7a-84cd-4b93-a1e4-a6e98502eb0f req-7ee91a4f-59ac-4ba8-a204-49a61c273202 service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] No waiting events found dispatching network-vif-plugged-4a9242b0-4afa-443e-9ccd-dc35313fc59a {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 872.607537] env[70013]: WARNING nova.compute.manager [req-0debad7a-84cd-4b93-a1e4-a6e98502eb0f req-7ee91a4f-59ac-4ba8-a204-49a61c273202 service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Received unexpected event network-vif-plugged-4a9242b0-4afa-443e-9ccd-dc35313fc59a for instance with vm_state building and task_state spawning. [ 872.639379] env[70013]: DEBUG nova.network.neutron [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Successfully updated port: 4a9242b0-4afa-443e-9ccd-dc35313fc59a {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.693192] env[70013]: DEBUG oslo_vmware.api [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Task: {'id': task-4230937, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178641} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.693458] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.693650] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 872.693812] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 872.694030] env[70013]: INFO nova.compute.manager [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Took 1.13 seconds to destroy the instance on the hypervisor. [ 872.694233] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 872.696921] env[70013]: DEBUG nova.compute.manager [-] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 872.697043] env[70013]: DEBUG nova.network.neutron [-] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 872.698809] env[70013]: INFO nova.compute.manager [-] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Took 1.42 seconds to deallocate network for instance. [ 872.777115] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230938, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.001859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.057723] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a33cf37-f088-4ca4-9385-a968a5c908ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.068197] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200f8f32-2301-4d12-a747-813f4fbbc593 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.104407] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000db534-03f0-49e7-b8dc-99d1c0e536a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.114118] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef4b7fa-ff3f-4d51-9981-eaa62716fb95 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.118774] env[70013]: DEBUG nova.network.neutron [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updated VIF entry in instance network info cache for port 95450791-e5ab-4650-be74-3d6299da9718. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.119152] env[70013]: DEBUG nova.network.neutron [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updating instance_info_cache with network_info: [{"id": "95450791-e5ab-4650-be74-3d6299da9718", "address": "fa:16:3e:39:92:a5", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95450791-e5", "ovs_interfaceid": "95450791-e5ab-4650-be74-3d6299da9718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.134018] env[70013]: DEBUG nova.compute.provider_tree [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.145174] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.145356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.145685] env[70013]: DEBUG nova.network.neutron [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.206363] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.270647] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230938, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.623189] env[70013]: DEBUG oslo_concurrency.lockutils [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] Releasing lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 873.623496] env[70013]: DEBUG nova.compute.manager [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Received event network-vif-deleted-b402fce4-7392-4139-8de5-7620820e0814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 873.623549] env[70013]: INFO nova.compute.manager [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Neutron deleted interface b402fce4-7392-4139-8de5-7620820e0814; detaching it from the instance and deleting it from the info cache [ 873.623764] env[70013]: DEBUG nova.network.neutron [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.635018] env[70013]: DEBUG nova.scheduler.client.report [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 873.662857] env[70013]: DEBUG nova.network.neutron [-] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.686920] env[70013]: DEBUG nova.network.neutron [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.774945] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230938, 'name': CreateVM_Task, 'duration_secs': 1.407314} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.775115] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.776153] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.776365] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.776769] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 873.777363] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-103a427a-80a0-497a-bf87-78def53b4e93 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.785985] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 873.785985] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525e5425-c618-6c28-ae27-412dab4bd109" [ 873.785985] env[70013]: _type = "Task" [ 873.785985] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.800519] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525e5425-c618-6c28-ae27-412dab4bd109, 'name': SearchDatastore_Task, 'duration_secs': 0.010359} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.802044] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 873.802844] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.802844] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.802844] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.803040] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.803286] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9359ede8-9fbc-4c30-8035-0929e5f43dfc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.815687] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.816310] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.816726] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cef9975-0634-404a-9426-7e7e71bdf30c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.824901] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 873.824901] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52048b80-400b-9802-55db-afb3e5cb9629" [ 873.824901] env[70013]: _type = "Task" [ 873.824901] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.839607] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52048b80-400b-9802-55db-afb3e5cb9629, 'name': SearchDatastore_Task, 'duration_secs': 0.010271} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.841499] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-369b4772-eafa-4be6-b96a-7e5b3b3ef6a3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.848120] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 873.848120] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52db0c5f-2a80-329c-6ccb-bc6205b9d3f1" [ 873.848120] env[70013]: _type = "Task" [ 873.848120] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.859927] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52db0c5f-2a80-329c-6ccb-bc6205b9d3f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.910696] env[70013]: DEBUG nova.network.neutron [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Updating instance_info_cache with network_info: [{"id": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "address": "fa:16:3e:01:6f:a2", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a9242b0-4a", "ovs_interfaceid": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.025423] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.025423] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.126526] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-214f46fc-4dd9-4fbf-9e49-da3c526883f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.137306] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909fb7bb-b02d-4fd9-bafb-055f85d52040 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.151096] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.184s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.151494] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.953s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.153479] env[70013]: INFO nova.compute.claims [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.169992] env[70013]: INFO nova.compute.manager [-] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Took 1.47 seconds to deallocate network for instance. [ 874.180171] env[70013]: DEBUG nova.compute.manager [req-9c1f435d-232d-4156-83a4-605eca7e65bd req-871d76d8-da73-4606-8062-72c14b347c7f service nova] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Detach interface failed, port_id=b402fce4-7392-4139-8de5-7620820e0814, reason: Instance 3b29c99a-0d55-40b1-a155-199b1ced2146 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 874.192046] env[70013]: INFO nova.scheduler.client.report [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Deleted allocations for instance 02ef7875-7ec9-4409-aaa5-71ed669f3780 [ 874.359775] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52db0c5f-2a80-329c-6ccb-bc6205b9d3f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010331} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.360099] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.360355] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] a53389ce-d0c2-48ae-83e3-ee58fbdbfb01/a53389ce-d0c2-48ae-83e3-ee58fbdbfb01.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.360626] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76208f87-cb34-46c9-bfb2-3b153edcc6c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.369336] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 874.369336] env[70013]: value = "task-4230939" [ 874.369336] env[70013]: _type = "Task" [ 874.369336] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.381273] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.414747] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.415163] env[70013]: DEBUG nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Instance network_info: |[{"id": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "address": "fa:16:3e:01:6f:a2", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a9242b0-4a", "ovs_interfaceid": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 874.415642] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:6f:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c13fd8bc-e797-42fe-94ed-6370d3467a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a9242b0-4afa-443e-9ccd-dc35313fc59a', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.424428] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Creating folder: Project (208376e0d950450cb2ce841e2d049cd2). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 874.424732] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-902f7a8f-3919-4a9d-9986-c5a814b6b25d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.440036] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Created folder: Project (208376e0d950450cb2ce841e2d049cd2) in parent group-v836999. [ 874.440036] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Creating folder: Instances. Parent ref: group-v837123. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 874.440036] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7044b0d4-9f42-4940-80c6-76f62dccd2a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.453483] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Created folder: Instances in parent group-v837123. [ 874.454102] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 874.454102] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 874.454313] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-caee24ba-a0a4-4166-aa0e-f44071503473 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.485988] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.485988] env[70013]: value = "task-4230942" [ 874.485988] env[70013]: _type = "Task" [ 874.485988] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.497645] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230942, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.611599] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.611967] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.612519] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.613017] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.613017] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.616419] env[70013]: INFO nova.compute.manager [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Terminating instance [ 874.667286] env[70013]: DEBUG nova.compute.manager [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Received event network-changed-4a9242b0-4afa-443e-9ccd-dc35313fc59a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 874.667684] env[70013]: DEBUG nova.compute.manager [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Refreshing instance network info cache due to event network-changed-4a9242b0-4afa-443e-9ccd-dc35313fc59a. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 874.668086] env[70013]: DEBUG oslo_concurrency.lockutils [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] Acquiring lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.668821] env[70013]: DEBUG oslo_concurrency.lockutils [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] Acquired lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.668821] env[70013]: DEBUG nova.network.neutron [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Refreshing network info cache for port 4a9242b0-4afa-443e-9ccd-dc35313fc59a {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.686454] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.703434] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4271f021-9ea8-4a1f-bdeb-ceebd68d150c tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "02ef7875-7ec9-4409-aaa5-71ed669f3780" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.539s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.882426] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230939, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.997104] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230942, 'name': CreateVM_Task, 'duration_secs': 0.478804} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.997292] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 874.998109] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.999264] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.999264] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 874.999264] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98a31812-3475-475e-b538-f19db15632cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.004288] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 875.004288] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200cd1e-8599-ee13-aa78-5eadf96f1733" [ 875.004288] env[70013]: _type = "Task" [ 875.004288] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.016508] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200cd1e-8599-ee13-aa78-5eadf96f1733, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.121609] env[70013]: DEBUG nova.compute.manager [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 875.121861] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.122797] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39e9dcd-f597-4b3a-aef5-21fabcadc66b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.130743] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 875.130963] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-002c8b72-03cd-4947-aea8-0ee469da46eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.137917] env[70013]: DEBUG oslo_vmware.api [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 875.137917] env[70013]: value = "task-4230943" [ 875.137917] env[70013]: _type = "Task" [ 875.137917] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.146421] env[70013]: DEBUG oslo_vmware.api [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230943, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.392278] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521428} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.395684] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] a53389ce-d0c2-48ae-83e3-ee58fbdbfb01/a53389ce-d0c2-48ae-83e3-ee58fbdbfb01.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.395684] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.396948] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b990a074-1a4a-425f-9777-ba881ac3ad34 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.405410] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 875.405410] env[70013]: value = "task-4230944" [ 875.405410] env[70013]: _type = "Task" [ 875.405410] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.421878] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.514297] env[70013]: DEBUG nova.network.neutron [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Updated VIF entry in instance network info cache for port 4a9242b0-4afa-443e-9ccd-dc35313fc59a. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 875.514297] env[70013]: DEBUG nova.network.neutron [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Updating instance_info_cache with network_info: [{"id": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "address": "fa:16:3e:01:6f:a2", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a9242b0-4a", "ovs_interfaceid": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.526440] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200cd1e-8599-ee13-aa78-5eadf96f1733, 'name': SearchDatastore_Task, 'duration_secs': 0.009641} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.530035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.530428] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.530639] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.530803] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 875.531029] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.531496] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6dc2f519-98d5-4085-805e-05aaead6dfdb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.542430] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.542620] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.546983] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3727802-4e60-446a-bafa-53a392b9f748 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.552334] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 875.552334] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fb2373-26a9-f78f-69d4-0293516851b4" [ 875.552334] env[70013]: _type = "Task" [ 875.552334] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.564207] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fb2373-26a9-f78f-69d4-0293516851b4, 'name': SearchDatastore_Task, 'duration_secs': 0.011452} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.565885] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce8ac899-69e1-414f-b2ac-86b575b7f97f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.573982] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 875.573982] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b4b6c-dd71-2e5d-5388-bc7395674fa2" [ 875.573982] env[70013]: _type = "Task" [ 875.573982] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.582725] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b4b6c-dd71-2e5d-5388-bc7395674fa2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.651981] env[70013]: DEBUG oslo_vmware.api [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230943, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.776063] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffdd175-1e98-46fc-9223-5ae15487fce8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.784366] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d677e878-ad5c-462b-bfc6-a1d8c1500040 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.817868] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a3cf6a-d907-4e53-be1d-a2a3ac1da8b7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.826235] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8451e673-a20d-4197-a5d2-e87c117f5633 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.842509] env[70013]: DEBUG nova.compute.provider_tree [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.916980] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075666} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.917336] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.918531] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76f76c7-cee0-4c2c-8d53-0a158973f1c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.946268] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] a53389ce-d0c2-48ae-83e3-ee58fbdbfb01/a53389ce-d0c2-48ae-83e3-ee58fbdbfb01.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.946658] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3af4ae19-8cbd-42f6-a5ac-c6ce6da1f864 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.975243] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 875.975243] env[70013]: value = "task-4230945" [ 875.975243] env[70013]: _type = "Task" [ 875.975243] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.984099] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230945, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.017478] env[70013]: DEBUG oslo_concurrency.lockutils [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] Releasing lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 876.018073] env[70013]: DEBUG nova.compute.manager [req-28c1b758-a127-423f-8596-4007669c1e71 req-96abd027-b7ed-4f80-9d37-e96d606be31b service nova] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Received event network-vif-deleted-3b0bd58d-e108-46ac-8fc1-5eeda2eb17e2 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 876.086489] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b4b6c-dd71-2e5d-5388-bc7395674fa2, 'name': SearchDatastore_Task, 'duration_secs': 0.010268} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.086928] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 876.088253] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/789fffd9-6725-4bf6-9144-dd603b0a521f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 876.088253] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c2865f1-f940-482d-8e5d-94dab0ed5a4c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.094843] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 876.094843] env[70013]: value = "task-4230946" [ 876.094843] env[70013]: _type = "Task" [ 876.094843] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.105221] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230946, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.152398] env[70013]: DEBUG oslo_vmware.api [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230943, 'name': PowerOffVM_Task, 'duration_secs': 0.985948} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.152833] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.152833] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.153206] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae84417b-fa6e-46fe-8db1-dfb215149f27 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.236712] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.236712] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.236712] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Deleting the datastore file [datastore1] 0b0b2769-74ba-4ba2-9e8b-148cc2976321 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.236712] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f5dfe23-4080-4a0c-a50b-31c765295b09 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.246346] env[70013]: DEBUG oslo_vmware.api [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for the task: (returnval){ [ 876.246346] env[70013]: value = "task-4230948" [ 876.246346] env[70013]: _type = "Task" [ 876.246346] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.260020] env[70013]: DEBUG oslo_vmware.api [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230948, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.304943] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 876.304943] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.350112] env[70013]: DEBUG nova.scheduler.client.report [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 876.486993] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230945, 'name': ReconfigVM_Task, 'duration_secs': 0.316639} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.488121] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Reconfigured VM instance instance-00000031 to attach disk [datastore2] a53389ce-d0c2-48ae-83e3-ee58fbdbfb01/a53389ce-d0c2-48ae-83e3-ee58fbdbfb01.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.489108] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10b6171a-8b7f-404b-8641-e9d0886c2df8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.500763] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 876.500763] env[70013]: value = "task-4230949" [ 876.500763] env[70013]: _type = "Task" [ 876.500763] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.519740] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230949, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.608395] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230946, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.758777] env[70013]: DEBUG oslo_vmware.api [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Task: {'id': task-4230948, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.448972} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.759036] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.763023] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 876.763023] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 876.763023] env[70013]: INFO nova.compute.manager [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Took 1.64 seconds to destroy the instance on the hypervisor. [ 876.763023] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 876.763023] env[70013]: DEBUG nova.compute.manager [-] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 876.763023] env[70013]: DEBUG nova.network.neutron [-] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 876.856272] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.704s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.856637] env[70013]: DEBUG nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 876.859808] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.497s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.864694] env[70013]: DEBUG nova.objects.instance [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lazy-loading 'resources' on Instance uuid b1d844cc-dcdc-47e2-85c3-484d0862be60 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.012907] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230949, 'name': Rename_Task, 'duration_secs': 0.18575} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.013229] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.013483] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5a50a3f-1f91-4cb6-91ab-db6b642cd7f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.024645] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 877.024645] env[70013]: value = "task-4230950" [ 877.024645] env[70013]: _type = "Task" [ 877.024645] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.036999] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230950, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.110181] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230946, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538337} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.110806] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/789fffd9-6725-4bf6-9144-dd603b0a521f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.113718] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.114049] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20ac2bae-ca90-43d5-aea1-fe80971d1ddf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.123451] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 877.123451] env[70013]: value = "task-4230951" [ 877.123451] env[70013]: _type = "Task" [ 877.123451] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.132901] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230951, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.371303] env[70013]: DEBUG nova.compute.utils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 877.380642] env[70013]: DEBUG nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 877.381053] env[70013]: DEBUG nova.network.neutron [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 877.428936] env[70013]: DEBUG nova.compute.manager [req-1fcd25e4-bb0d-49e9-babf-2076705cfb74 req-55a753ce-7b83-4024-a64d-7d072afc1c0f service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Received event network-vif-deleted-f3add360-018f-4a8b-9a58-1a47a83f2bc4 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 877.429173] env[70013]: INFO nova.compute.manager [req-1fcd25e4-bb0d-49e9-babf-2076705cfb74 req-55a753ce-7b83-4024-a64d-7d072afc1c0f service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Neutron deleted interface f3add360-018f-4a8b-9a58-1a47a83f2bc4; detaching it from the instance and deleting it from the info cache [ 877.429414] env[70013]: DEBUG nova.network.neutron [req-1fcd25e4-bb0d-49e9-babf-2076705cfb74 req-55a753ce-7b83-4024-a64d-7d072afc1c0f service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.450021] env[70013]: DEBUG nova.policy [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a8d904839254e55bd56cd8b87b93711', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '208376e0d950450cb2ce841e2d049cd2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 877.539852] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230950, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.634475] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230951, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074887} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.634767] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.635805] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ed3068-c9fe-44f5-811d-c984c756d10a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.663958] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/789fffd9-6725-4bf6-9144-dd603b0a521f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.667728] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-621a40c0-03f9-4b59-8302-9729b2ccc3b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.689705] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 877.689705] env[70013]: value = "task-4230952" [ 877.689705] env[70013]: _type = "Task" [ 877.689705] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.702276] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230952, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.880852] env[70013]: DEBUG nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 877.895483] env[70013]: DEBUG nova.network.neutron [-] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.932298] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3fe41861-9d05-4dec-a958-1856170143aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.944413] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6799efb-f2b6-472e-9ad1-5436c808ce17 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.983628] env[70013]: DEBUG nova.compute.manager [req-1fcd25e4-bb0d-49e9-babf-2076705cfb74 req-55a753ce-7b83-4024-a64d-7d072afc1c0f service nova] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Detach interface failed, port_id=f3add360-018f-4a8b-9a58-1a47a83f2bc4, reason: Instance 0b0b2769-74ba-4ba2-9e8b-148cc2976321 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 878.028122] env[70013]: DEBUG nova.network.neutron [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Successfully created port: 25ae3868-dd45-4f4c-8b91-809dcb235f5c {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.040199] env[70013]: DEBUG oslo_vmware.api [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230950, 'name': PowerOnVM_Task, 'duration_secs': 0.762607} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.042101] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.042332] env[70013]: INFO nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Took 9.07 seconds to spawn the instance on the hypervisor. [ 878.048018] env[70013]: DEBUG nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 878.048018] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb8ae0f-2173-4b61-9118-50fce8929a1d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.048018] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edcc7f52-174b-4d33-bd64-3a3846194b34 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.064751] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f567e3b-0f58-427b-9b70-2c290d9a34bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.103184] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd2b06c-84c0-489b-9e51-aa1f3ec69b90 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.113778] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f55ba56-efee-4150-b5bb-0123a534863c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.130042] env[70013]: DEBUG nova.compute.provider_tree [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.200753] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230952, 'name': ReconfigVM_Task, 'duration_secs': 0.339099} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.201078] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/789fffd9-6725-4bf6-9144-dd603b0a521f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.201708] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ddbcbb2-dd40-4b82-9764-4cf4a1bda6d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.209422] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 878.209422] env[70013]: value = "task-4230953" [ 878.209422] env[70013]: _type = "Task" [ 878.209422] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.222023] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230953, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.398805] env[70013]: INFO nova.compute.manager [-] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Took 1.64 seconds to deallocate network for instance. [ 878.579587] env[70013]: INFO nova.compute.manager [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Took 43.79 seconds to build instance. [ 878.635595] env[70013]: DEBUG nova.scheduler.client.report [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 878.721271] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230953, 'name': Rename_Task, 'duration_secs': 0.185805} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.721733] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.722138] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12b72fa1-e76a-4fde-892d-8e7db9950fec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.730756] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 878.730756] env[70013]: value = "task-4230954" [ 878.730756] env[70013]: _type = "Task" [ 878.730756] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.739517] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.891621] env[70013]: DEBUG nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 878.916388] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.927421] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 878.927748] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.927992] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 878.928424] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.928511] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 878.928734] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 878.929053] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 878.929264] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 878.929759] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 878.929759] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 878.929996] env[70013]: DEBUG nova.virt.hardware [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 878.932689] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3e4eb9-c604-45f9-b56c-61fc7d685f96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.945619] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12789a91-7f39-4b22-9cf4-331c7ecf5324 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.082586] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90390653-83ad-42f4-a007-d6889770846e tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.886s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.140688] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.281s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.143272] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.087s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.143556] env[70013]: DEBUG nova.objects.instance [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'resources' on Instance uuid 9a0ecd6c-5853-446d-8f41-b2ee51e5259a {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.250073] env[70013]: DEBUG oslo_vmware.api [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230954, 'name': PowerOnVM_Task, 'duration_secs': 0.512815} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.250489] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.250777] env[70013]: INFO nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Took 7.51 seconds to spawn the instance on the hypervisor. [ 879.251055] env[70013]: DEBUG nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.252148] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2049c8-a0ac-473e-816f-a3bf530b2fd4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.586816] env[70013]: DEBUG nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 879.676812] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e56b133-b1a6-40a0-b58c-b4588c81849b tempest-ListServersNegativeTestJSON-1085704939 tempest-ListServersNegativeTestJSON-1085704939-project-member] Lock "b1d844cc-dcdc-47e2-85c3-484d0862be60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.504s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.730096] env[70013]: DEBUG nova.network.neutron [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Successfully updated port: 25ae3868-dd45-4f4c-8b91-809dcb235f5c {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.774937] env[70013]: INFO nova.compute.manager [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Took 43.69 seconds to build instance. [ 880.110802] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.123842] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df1d5a1-55a1-42a3-ae98-5e51cd83d1a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.133050] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28191a3-48e4-47fc-bfc1-c2fe6b181398 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.167720] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1068055-a7c0-4fb9-bf3c-d81d039272e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.178075] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fad167c-89c5-43b5-b69d-9ef5accbcb8e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.197671] env[70013]: DEBUG nova.compute.provider_tree [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.235950] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.236369] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.236458] env[70013]: DEBUG nova.network.neutron [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.282460] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e950401c-b3f3-44d6-aa28-9957a82fec86 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.242s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.429386] env[70013]: DEBUG nova.compute.manager [req-ec2cb380-a697-4a42-94c8-31cd83e17d0e req-69debac1-80f8-4678-ba4c-0ad6adb79bb2 service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Received event network-vif-plugged-25ae3868-dd45-4f4c-8b91-809dcb235f5c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 880.429618] env[70013]: DEBUG oslo_concurrency.lockutils [req-ec2cb380-a697-4a42-94c8-31cd83e17d0e req-69debac1-80f8-4678-ba4c-0ad6adb79bb2 service nova] Acquiring lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.429842] env[70013]: DEBUG oslo_concurrency.lockutils [req-ec2cb380-a697-4a42-94c8-31cd83e17d0e req-69debac1-80f8-4678-ba4c-0ad6adb79bb2 service nova] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.430042] env[70013]: DEBUG oslo_concurrency.lockutils [req-ec2cb380-a697-4a42-94c8-31cd83e17d0e req-69debac1-80f8-4678-ba4c-0ad6adb79bb2 service nova] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.430224] env[70013]: DEBUG nova.compute.manager [req-ec2cb380-a697-4a42-94c8-31cd83e17d0e req-69debac1-80f8-4678-ba4c-0ad6adb79bb2 service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] No waiting events found dispatching network-vif-plugged-25ae3868-dd45-4f4c-8b91-809dcb235f5c {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 880.430594] env[70013]: WARNING nova.compute.manager [req-ec2cb380-a697-4a42-94c8-31cd83e17d0e req-69debac1-80f8-4678-ba4c-0ad6adb79bb2 service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Received unexpected event network-vif-plugged-25ae3868-dd45-4f4c-8b91-809dcb235f5c for instance with vm_state building and task_state spawning. [ 880.555072] env[70013]: DEBUG nova.compute.manager [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Received event network-changed-95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 880.555129] env[70013]: DEBUG nova.compute.manager [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Refreshing instance network info cache due to event network-changed-95450791-e5ab-4650-be74-3d6299da9718. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 880.555327] env[70013]: DEBUG oslo_concurrency.lockutils [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] Acquiring lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.555635] env[70013]: DEBUG oslo_concurrency.lockutils [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] Acquired lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.555857] env[70013]: DEBUG nova.network.neutron [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Refreshing network info cache for port 95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.621136] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "ab091ef4-9828-438f-8e95-4370bfbd34de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.621301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.701703] env[70013]: DEBUG nova.scheduler.client.report [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 880.779644] env[70013]: DEBUG nova.network.neutron [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.786837] env[70013]: DEBUG nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 880.938583] env[70013]: DEBUG nova.network.neutron [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Updating instance_info_cache with network_info: [{"id": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "address": "fa:16:3e:cd:a2:20", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ae3868-dd", "ovs_interfaceid": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.207539] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.064s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 881.210630] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.129s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 881.213609] env[70013]: INFO nova.compute.claims [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.247602] env[70013]: DEBUG oslo_concurrency.lockutils [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.247874] env[70013]: DEBUG oslo_concurrency.lockutils [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 881.248240] env[70013]: INFO nova.compute.manager [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Rebooting instance [ 881.252068] env[70013]: INFO nova.scheduler.client.report [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted allocations for instance 9a0ecd6c-5853-446d-8f41-b2ee51e5259a [ 881.303246] env[70013]: DEBUG nova.network.neutron [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updated VIF entry in instance network info cache for port 95450791-e5ab-4650-be74-3d6299da9718. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.303607] env[70013]: DEBUG nova.network.neutron [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updating instance_info_cache with network_info: [{"id": "95450791-e5ab-4650-be74-3d6299da9718", "address": "fa:16:3e:39:92:a5", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95450791-e5", "ovs_interfaceid": "95450791-e5ab-4650-be74-3d6299da9718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.311179] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.447176] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.447176] env[70013]: DEBUG nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Instance network_info: |[{"id": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "address": "fa:16:3e:cd:a2:20", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ae3868-dd", "ovs_interfaceid": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 881.447176] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:a2:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c13fd8bc-e797-42fe-94ed-6370d3467a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25ae3868-dd45-4f4c-8b91-809dcb235f5c', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.457271] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 881.457916] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.458334] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ad3c04e-2493-4183-9f93-3490353baa52 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.487264] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.487264] env[70013]: value = "task-4230955" [ 881.487264] env[70013]: _type = "Task" [ 881.487264] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.495226] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230955, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.762558] env[70013]: DEBUG oslo_concurrency.lockutils [None req-995a6614-e604-4d06-9435-13abb3f05f88 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "9a0ecd6c-5853-446d-8f41-b2ee51e5259a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.344s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 881.781172] env[70013]: DEBUG oslo_concurrency.lockutils [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.812251] env[70013]: DEBUG oslo_concurrency.lockutils [req-6029253d-a5c3-4351-94d9-fc45e0f4586a req-26a9ba56-1348-4492-9ea2-ba928366447d service nova] Releasing lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.812251] env[70013]: DEBUG oslo_concurrency.lockutils [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquired lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.812251] env[70013]: DEBUG nova.network.neutron [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.998293] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230955, 'name': CreateVM_Task, 'duration_secs': 0.43656} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.998561] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.999304] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.999481] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.999809] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 882.000170] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd750e43-bf8e-484a-a067-4d8c6618cf8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.006574] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 882.006574] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e567b4-ba84-efe1-2d45-08bb3c2003f0" [ 882.006574] env[70013]: _type = "Task" [ 882.006574] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.015560] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e567b4-ba84-efe1-2d45-08bb3c2003f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.506992] env[70013]: DEBUG nova.compute.manager [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Received event network-changed-25ae3868-dd45-4f4c-8b91-809dcb235f5c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 882.507197] env[70013]: DEBUG nova.compute.manager [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Refreshing instance network info cache due to event network-changed-25ae3868-dd45-4f4c-8b91-809dcb235f5c. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 882.507569] env[70013]: DEBUG oslo_concurrency.lockutils [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] Acquiring lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.507569] env[70013]: DEBUG oslo_concurrency.lockutils [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] Acquired lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 882.507737] env[70013]: DEBUG nova.network.neutron [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Refreshing network info cache for port 25ae3868-dd45-4f4c-8b91-809dcb235f5c {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.524393] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e567b4-ba84-efe1-2d45-08bb3c2003f0, 'name': SearchDatastore_Task, 'duration_secs': 0.012884} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.528343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.528608] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.528849] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.528995] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 882.529199] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.530061] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2899da47-2bc6-41b5-9180-b8ae6ea0d12a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.540920] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.541163] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.544816] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c92a5710-5636-4904-aa22-94fc54dc3cd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.554069] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 882.554069] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5217a79b-e51b-3455-48f5-2f61d68952dd" [ 882.554069] env[70013]: _type = "Task" [ 882.554069] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.564771] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5217a79b-e51b-3455-48f5-2f61d68952dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.740223] env[70013]: DEBUG nova.network.neutron [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updating instance_info_cache with network_info: [{"id": "95450791-e5ab-4650-be74-3d6299da9718", "address": "fa:16:3e:39:92:a5", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95450791-e5", "ovs_interfaceid": "95450791-e5ab-4650-be74-3d6299da9718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.758229] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2abf783-1c6b-4efa-82b5-cd9b59b65b39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.769019] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c1db08-ac3f-4ad8-9339-652ec2186f3c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.805088] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0dc3b9-46a5-40a7-b35b-e1c20d033ee6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.814265] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06bac67-f6d4-47ae-9b67-5a3217aaa497 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.831053] env[70013]: DEBUG nova.compute.provider_tree [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.066531] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5217a79b-e51b-3455-48f5-2f61d68952dd, 'name': SearchDatastore_Task, 'duration_secs': 0.01138} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.067378] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-598bc02d-2e75-4fad-b922-c21a08be0530 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.076245] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 883.076245] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529c5f16-225a-e98e-f2d1-d1ede105fa4c" [ 883.076245] env[70013]: _type = "Task" [ 883.076245] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.086359] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529c5f16-225a-e98e-f2d1-d1ede105fa4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.229352] env[70013]: DEBUG nova.network.neutron [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Updated VIF entry in instance network info cache for port 25ae3868-dd45-4f4c-8b91-809dcb235f5c. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 883.229998] env[70013]: DEBUG nova.network.neutron [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Updating instance_info_cache with network_info: [{"id": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "address": "fa:16:3e:cd:a2:20", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ae3868-dd", "ovs_interfaceid": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.248229] env[70013]: DEBUG oslo_concurrency.lockutils [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Releasing lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.335136] env[70013]: DEBUG nova.scheduler.client.report [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 883.588646] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529c5f16-225a-e98e-f2d1-d1ede105fa4c, 'name': SearchDatastore_Task, 'duration_secs': 0.011985} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.588973] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.589341] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/ef64a05a-b514-4c35-81d3-664ae1ad3ff1.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.589629] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-071c14af-4490-4e23-a0b9-e4bee6c93a45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.598415] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 883.598415] env[70013]: value = "task-4230956" [ 883.598415] env[70013]: _type = "Task" [ 883.598415] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.608819] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230956, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.733681] env[70013]: DEBUG oslo_concurrency.lockutils [req-71fb50c0-44b1-42cc-8e64-46a9e464da30 req-53dadae2-eb25-446e-9e21-8dd77dcd4fcf service nova] Releasing lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.756667] env[70013]: DEBUG nova.compute.manager [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 883.757624] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e8c5d4-1c9f-4de4-9e81-67f07c3b9b4d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.845512] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.846319] env[70013]: DEBUG nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 883.849516] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.709s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.852298] env[70013]: INFO nova.compute.claims [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.113949] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230956, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.359021] env[70013]: DEBUG nova.compute.utils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 884.361209] env[70013]: DEBUG nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 884.361413] env[70013]: DEBUG nova.network.neutron [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 884.414413] env[70013]: DEBUG nova.policy [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72233f434f7f40e08ba4e4740c17499f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f16290bcd228490fbdb41439404d61d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 884.609139] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230956, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526142} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.609402] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/ef64a05a-b514-4c35-81d3-664ae1ad3ff1.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 884.609917] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.609917] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc81da67-1769-4a1e-863b-860dfe9c79b1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.617145] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 884.617145] env[70013]: value = "task-4230957" [ 884.617145] env[70013]: _type = "Task" [ 884.617145] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.628435] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230957, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.779279] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e99c06-8071-46dd-afc4-5f508ae66c1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.794194] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Doing hard reboot of VM {{(pid=70013) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 884.794194] env[70013]: DEBUG nova.network.neutron [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Successfully created port: 04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.796663] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-46094824-76e4-420d-b7f4-90769da2fc58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.806186] env[70013]: DEBUG oslo_vmware.api [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 884.806186] env[70013]: value = "task-4230958" [ 884.806186] env[70013]: _type = "Task" [ 884.806186] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.813553] env[70013]: DEBUG oslo_vmware.api [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230958, 'name': ResetVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.870636] env[70013]: DEBUG nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 885.046922] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.047337] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.128090] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230957, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.181013} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.130858] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.131877] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299d947d-4941-4618-bfa8-64bea714df00 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.158841] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/ef64a05a-b514-4c35-81d3-664ae1ad3ff1.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.162064] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb27e73a-321a-4814-9eb3-8fe6ae82f4bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.185562] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 885.185562] env[70013]: value = "task-4230959" [ 885.185562] env[70013]: _type = "Task" [ 885.185562] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.198781] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230959, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.319665] env[70013]: DEBUG oslo_vmware.api [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230958, 'name': ResetVM_Task, 'duration_secs': 0.127581} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.322580] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Did hard reboot of VM {{(pid=70013) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 885.322799] env[70013]: DEBUG nova.compute.manager [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 885.324163] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7308d2-9ec5-4e33-b2b6-3c0d0dc44813 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.409415] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bee511-7b54-417e-8bc1-febbc337acab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.418545] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b659e2d3-ec41-49b2-ae75-9aa3be70a8ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.450423] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aded1317-b6f1-4a7b-84c8-f4c347df438c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.460789] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a21be7-ed98-4d73-807e-61bcb9d4827e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.477036] env[70013]: DEBUG nova.compute.provider_tree [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.697192] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230959, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.841807] env[70013]: DEBUG oslo_concurrency.lockutils [None req-34011212-d867-497e-938f-155081c8aba4 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.594s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.887647] env[70013]: DEBUG nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 885.918877] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 885.919553] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.919553] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 885.919553] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.919730] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 885.919962] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 885.920081] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 885.920246] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 885.920416] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 885.920692] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 885.920799] env[70013]: DEBUG nova.virt.hardware [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 885.921727] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c303f4aa-a5ac-41b0-80ab-0d27e93e0ee7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.931302] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df540a3-4289-41e9-a994-43a0f9764aec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.980762] env[70013]: DEBUG nova.scheduler.client.report [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 886.199983] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230959, 'name': ReconfigVM_Task, 'duration_secs': 1.005235} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.200404] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Reconfigured VM instance instance-00000033 to attach disk [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/ef64a05a-b514-4c35-81d3-664ae1ad3ff1.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.201202] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de19e3ef-d595-4edc-93a5-22f632a02c67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.209279] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 886.209279] env[70013]: value = "task-4230960" [ 886.209279] env[70013]: _type = "Task" [ 886.209279] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.218956] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230960, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.375215] env[70013]: DEBUG nova.network.neutron [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Successfully updated port: 04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.486982] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.487548] env[70013]: DEBUG nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 886.490673] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.778s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 886.494146] env[70013]: DEBUG nova.objects.instance [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lazy-loading 'resources' on Instance uuid b23b4e67-67a6-4117-9a62-3c0dcd25e828 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.723138] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230960, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.879811] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.880053] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquired lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.880258] env[70013]: DEBUG nova.network.neutron [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.962849] env[70013]: DEBUG nova.compute.manager [req-a81ac065-bb93-4ab0-ad84-ccd62cc3635a req-4bb70f86-4d80-4c43-9efe-669001da35bd service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Received event network-vif-plugged-04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 886.963204] env[70013]: DEBUG oslo_concurrency.lockutils [req-a81ac065-bb93-4ab0-ad84-ccd62cc3635a req-4bb70f86-4d80-4c43-9efe-669001da35bd service nova] Acquiring lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 886.963352] env[70013]: DEBUG oslo_concurrency.lockutils [req-a81ac065-bb93-4ab0-ad84-ccd62cc3635a req-4bb70f86-4d80-4c43-9efe-669001da35bd service nova] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 886.963514] env[70013]: DEBUG oslo_concurrency.lockutils [req-a81ac065-bb93-4ab0-ad84-ccd62cc3635a req-4bb70f86-4d80-4c43-9efe-669001da35bd service nova] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.963637] env[70013]: DEBUG nova.compute.manager [req-a81ac065-bb93-4ab0-ad84-ccd62cc3635a req-4bb70f86-4d80-4c43-9efe-669001da35bd service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] No waiting events found dispatching network-vif-plugged-04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 886.963767] env[70013]: WARNING nova.compute.manager [req-a81ac065-bb93-4ab0-ad84-ccd62cc3635a req-4bb70f86-4d80-4c43-9efe-669001da35bd service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Received unexpected event network-vif-plugged-04205eac-90e1-47c0-82c4-703b9258de68 for instance with vm_state building and task_state spawning. [ 886.995464] env[70013]: DEBUG nova.compute.utils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 887.000462] env[70013]: DEBUG nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 887.000799] env[70013]: DEBUG nova.network.neutron [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 887.073893] env[70013]: DEBUG nova.policy [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '377273afa56e450aa2c79fafc9b33626', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b9b70a7428848b7b1f0c618fd9edffc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 887.227392] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230960, 'name': Rename_Task, 'duration_secs': 0.97889} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.230032] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 887.230032] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c020ba9a-1a83-4a67-8b92-6eeb6bddff4c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.235831] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 887.235831] env[70013]: value = "task-4230961" [ 887.235831] env[70013]: _type = "Task" [ 887.235831] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.245251] env[70013]: DEBUG nova.compute.manager [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Received event network-changed-95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 887.245759] env[70013]: DEBUG nova.compute.manager [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Refreshing instance network info cache due to event network-changed-95450791-e5ab-4650-be74-3d6299da9718. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 887.245926] env[70013]: DEBUG oslo_concurrency.lockutils [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] Acquiring lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.246099] env[70013]: DEBUG oslo_concurrency.lockutils [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] Acquired lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.246351] env[70013]: DEBUG nova.network.neutron [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Refreshing network info cache for port 95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.256851] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230961, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.431304] env[70013]: DEBUG nova.network.neutron [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.501047] env[70013]: DEBUG nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 887.504440] env[70013]: DEBUG nova.network.neutron [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Successfully created port: b1d1b3a0-6686-4faf-981e-6772c3a821dd {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.619094] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7401d9d-d1a2-4d9f-bd54-50e8ad811191 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.632045] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559a91ea-1e44-487b-8cfa-9200dbb0745e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.662594] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bcd2697-0a0b-4a2f-a150-827838eec093 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.671773] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07fe815-3a70-40d0-b33f-a1f0f0faceb1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.686262] env[70013]: DEBUG nova.compute.provider_tree [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.697670] env[70013]: DEBUG nova.network.neutron [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Updating instance_info_cache with network_info: [{"id": "04205eac-90e1-47c0-82c4-703b9258de68", "address": "fa:16:3e:8e:33:9e", "network": {"id": "1d6a6584-952d-4f79-ba84-877f008c3ba9", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-840362347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f16290bcd228490fbdb41439404d61d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04205eac-90", "ovs_interfaceid": "04205eac-90e1-47c0-82c4-703b9258de68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.746946] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230961, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.051692] env[70013]: DEBUG nova.network.neutron [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updated VIF entry in instance network info cache for port 95450791-e5ab-4650-be74-3d6299da9718. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.052038] env[70013]: DEBUG nova.network.neutron [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updating instance_info_cache with network_info: [{"id": "95450791-e5ab-4650-be74-3d6299da9718", "address": "fa:16:3e:39:92:a5", "network": {"id": "0a69de94-3d78-4e55-9f62-94f5fc74a7a5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-178755679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85175ef2b004d5ea372f2dfe69549ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95450791-e5", "ovs_interfaceid": "95450791-e5ab-4650-be74-3d6299da9718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.195086] env[70013]: DEBUG nova.scheduler.client.report [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 888.205214] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Releasing lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.205214] env[70013]: DEBUG nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Instance network_info: |[{"id": "04205eac-90e1-47c0-82c4-703b9258de68", "address": "fa:16:3e:8e:33:9e", "network": {"id": "1d6a6584-952d-4f79-ba84-877f008c3ba9", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-840362347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f16290bcd228490fbdb41439404d61d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04205eac-90", "ovs_interfaceid": "04205eac-90e1-47c0-82c4-703b9258de68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 888.205214] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:33:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24144f5a-050a-4f1e-8d8c-774dc16dc791', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04205eac-90e1-47c0-82c4-703b9258de68', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.218017] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Creating folder: Project (f16290bcd228490fbdb41439404d61d9). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.218017] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8ca0d05-8d62-4af5-b710-c0832836cbca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.223578] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 888.223920] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.224203] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 888.224429] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.224644] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.227325] env[70013]: INFO nova.compute.manager [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Terminating instance [ 888.233517] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Created folder: Project (f16290bcd228490fbdb41439404d61d9) in parent group-v836999. [ 888.233914] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Creating folder: Instances. Parent ref: group-v837127. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.235056] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d89d74ee-8a60-4238-b8a9-1f89a98a1bbe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.248113] env[70013]: DEBUG oslo_vmware.api [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230961, 'name': PowerOnVM_Task, 'duration_secs': 0.667203} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.248475] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.248752] env[70013]: INFO nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Took 9.36 seconds to spawn the instance on the hypervisor. [ 888.249330] env[70013]: DEBUG nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 888.252297] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc99f72-01cd-4315-8be9-1c9b7b0fcd09 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.256336] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Created folder: Instances in parent group-v837127. [ 888.256658] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 888.257392] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.258096] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d6b8399-82ad-4735-aa75-d81cf6b06cc8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.286875] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.286875] env[70013]: value = "task-4230964" [ 888.286875] env[70013]: _type = "Task" [ 888.286875] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.296627] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230964, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.518349] env[70013]: DEBUG nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 888.556972] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 888.557366] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.557582] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 888.558334] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.558562] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 888.558748] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 888.559139] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 888.559361] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 888.559686] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 888.559961] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 888.560208] env[70013]: DEBUG nova.virt.hardware [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 888.561114] env[70013]: DEBUG oslo_concurrency.lockutils [req-71988b58-1177-40b1-8c10-7862bfb106f1 req-67405065-f5be-498a-9b4b-20ce25b5ce93 service nova] Releasing lock "refresh_cache-a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.562187] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c6429f-a65c-4798-b3fc-7234475be146 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.572341] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542b5101-9e87-4d1b-aa03-bf5ce51e4478 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.699939] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.209s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.707026] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.619s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.707026] env[70013]: INFO nova.compute.claims [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.727776] env[70013]: INFO nova.scheduler.client.report [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Deleted allocations for instance b23b4e67-67a6-4117-9a62-3c0dcd25e828 [ 888.734086] env[70013]: DEBUG nova.compute.manager [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 888.734566] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.735702] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d631d1a-b905-4d82-a3d2-af33b10aecf5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.744830] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.745189] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9f9bbc5-9bc0-4ef6-9f54-49b2e7775e33 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.753229] env[70013]: DEBUG oslo_vmware.api [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 888.753229] env[70013]: value = "task-4230965" [ 888.753229] env[70013]: _type = "Task" [ 888.753229] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.764441] env[70013]: DEBUG oslo_vmware.api [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230965, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.796688] env[70013]: INFO nova.compute.manager [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Took 46.61 seconds to build instance. [ 888.804716] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230964, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.012965] env[70013]: DEBUG nova.compute.manager [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Received event network-changed-04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 889.013336] env[70013]: DEBUG nova.compute.manager [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Refreshing instance network info cache due to event network-changed-04205eac-90e1-47c0-82c4-703b9258de68. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 889.013596] env[70013]: DEBUG oslo_concurrency.lockutils [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] Acquiring lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.013815] env[70013]: DEBUG oslo_concurrency.lockutils [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] Acquired lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 889.014052] env[70013]: DEBUG nova.network.neutron [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Refreshing network info cache for port 04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.112738] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Acquiring lock "55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.113227] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Lock "55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.244171] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7a7758f9-42cd-480a-b234-c949878378ca tempest-ServersWithSpecificFlavorTestJSON-1233962006 tempest-ServersWithSpecificFlavorTestJSON-1233962006-project-member] Lock "b23b4e67-67a6-4117-9a62-3c0dcd25e828" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.365s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.269025] env[70013]: DEBUG oslo_vmware.api [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230965, 'name': PowerOffVM_Task, 'duration_secs': 0.246836} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.269025] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.269025] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.269025] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a383caab-5e33-4123-9c71-bc08c87af21c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.299597] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba5bae4c-316a-47a9-9b55-7aac7eb3f0ba tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.985s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.305753] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230964, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.339162] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.339162] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.339162] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Deleting the datastore file [datastore2] a53389ce-d0c2-48ae-83e3-ee58fbdbfb01 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.339162] env[70013]: DEBUG nova.network.neutron [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Successfully updated port: b1d1b3a0-6686-4faf-981e-6772c3a821dd {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.340805] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca482f54-f5a3-4fc0-b389-7a44a94e03a0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.349809] env[70013]: DEBUG oslo_vmware.api [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 889.349809] env[70013]: value = "task-4230967" [ 889.349809] env[70013]: _type = "Task" [ 889.349809] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.360790] env[70013]: DEBUG oslo_vmware.api [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.808463] env[70013]: DEBUG nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 889.811737] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230964, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.847366] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "refresh_cache-0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.847466] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquired lock "refresh_cache-0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 889.849435] env[70013]: DEBUG nova.network.neutron [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.868688] env[70013]: DEBUG oslo_vmware.api [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4230967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139765} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.870350] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.871037] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.871037] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.871037] env[70013]: INFO nova.compute.manager [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Took 1.14 seconds to destroy the instance on the hypervisor. [ 889.871256] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 889.871888] env[70013]: DEBUG nova.compute.manager [-] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 889.871987] env[70013]: DEBUG nova.network.neutron [-] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.111607] env[70013]: DEBUG nova.network.neutron [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Updated VIF entry in instance network info cache for port 04205eac-90e1-47c0-82c4-703b9258de68. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.111767] env[70013]: DEBUG nova.network.neutron [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Updating instance_info_cache with network_info: [{"id": "04205eac-90e1-47c0-82c4-703b9258de68", "address": "fa:16:3e:8e:33:9e", "network": {"id": "1d6a6584-952d-4f79-ba84-877f008c3ba9", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-840362347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f16290bcd228490fbdb41439404d61d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04205eac-90", "ovs_interfaceid": "04205eac-90e1-47c0-82c4-703b9258de68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.309371] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230964, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.315016] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2afed13-0d3b-4e2c-9d19-5732c7d7d26a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.327936] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10a09ba-4b3d-4dee-9364-8cd2dd9571dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.333682] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.362916] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad88521-43c2-460e-8baf-77529ccce5a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.373849] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71478a4c-f545-472b-8722-4b99d5ab8626 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.390547] env[70013]: DEBUG nova.compute.provider_tree [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.398816] env[70013]: DEBUG nova.network.neutron [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.544356] env[70013]: DEBUG nova.network.neutron [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Updating instance_info_cache with network_info: [{"id": "b1d1b3a0-6686-4faf-981e-6772c3a821dd", "address": "fa:16:3e:a9:59:92", "network": {"id": "c639398a-1ce5-43bf-937a-a66b622e33f1", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-565920473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9b70a7428848b7b1f0c618fd9edffc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1d1b3a0-66", "ovs_interfaceid": "b1d1b3a0-6686-4faf-981e-6772c3a821dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.611540] env[70013]: INFO nova.compute.manager [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Rescuing [ 890.611823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.612011] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.612230] env[70013]: DEBUG nova.network.neutron [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.615303] env[70013]: DEBUG oslo_concurrency.lockutils [req-bf067176-2bbb-4de1-8c67-e0753841cc56 req-be735b0b-3dc0-4341-aa92-750e4b20acb4 service nova] Releasing lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.742777] env[70013]: DEBUG nova.network.neutron [-] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.805252] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230964, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.901745] env[70013]: DEBUG nova.scheduler.client.report [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 891.047739] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Releasing lock "refresh_cache-0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.048197] env[70013]: DEBUG nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Instance network_info: |[{"id": "b1d1b3a0-6686-4faf-981e-6772c3a821dd", "address": "fa:16:3e:a9:59:92", "network": {"id": "c639398a-1ce5-43bf-937a-a66b622e33f1", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-565920473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9b70a7428848b7b1f0c618fd9edffc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1d1b3a0-66", "ovs_interfaceid": "b1d1b3a0-6686-4faf-981e-6772c3a821dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 891.048699] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:59:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e0c77754-4085-434b-a3e8-d61be099ac67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1d1b3a0-6686-4faf-981e-6772c3a821dd', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.058606] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Creating folder: Project (0b9b70a7428848b7b1f0c618fd9edffc). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.058606] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-862b95b2-43ac-4b09-a1af-d611dabaa2de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.073487] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Created folder: Project (0b9b70a7428848b7b1f0c618fd9edffc) in parent group-v836999. [ 891.073710] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Creating folder: Instances. Parent ref: group-v837130. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.073975] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4841cf0-fd36-4aee-b1ae-2a6ee66d650a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.085779] env[70013]: DEBUG nova.compute.manager [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Received event network-vif-plugged-b1d1b3a0-6686-4faf-981e-6772c3a821dd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 891.085968] env[70013]: DEBUG oslo_concurrency.lockutils [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] Acquiring lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.086257] env[70013]: DEBUG oslo_concurrency.lockutils [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.086399] env[70013]: DEBUG oslo_concurrency.lockutils [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.086554] env[70013]: DEBUG nova.compute.manager [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] No waiting events found dispatching network-vif-plugged-b1d1b3a0-6686-4faf-981e-6772c3a821dd {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 891.086714] env[70013]: WARNING nova.compute.manager [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Received unexpected event network-vif-plugged-b1d1b3a0-6686-4faf-981e-6772c3a821dd for instance with vm_state building and task_state spawning. [ 891.087802] env[70013]: DEBUG nova.compute.manager [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Received event network-changed-b1d1b3a0-6686-4faf-981e-6772c3a821dd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 891.087802] env[70013]: DEBUG nova.compute.manager [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Refreshing instance network info cache due to event network-changed-b1d1b3a0-6686-4faf-981e-6772c3a821dd. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 891.087802] env[70013]: DEBUG oslo_concurrency.lockutils [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] Acquiring lock "refresh_cache-0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.087802] env[70013]: DEBUG oslo_concurrency.lockutils [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] Acquired lock "refresh_cache-0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.087802] env[70013]: DEBUG nova.network.neutron [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Refreshing network info cache for port b1d1b3a0-6686-4faf-981e-6772c3a821dd {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.091242] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Created folder: Instances in parent group-v837130. [ 891.091620] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 891.091914] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.092170] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a096556-50c3-4c8c-bdad-a5ff79855356 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.895990] env[70013]: INFO nova.compute.manager [-] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Took 2.02 seconds to deallocate network for instance. [ 891.897271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.193s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.897400] env[70013]: DEBUG nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 891.911041] env[70013]: DEBUG oslo_concurrency.lockutils [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.260s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.926024] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230964, 'name': CreateVM_Task, 'duration_secs': 3.560252} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.926024] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.926024] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.926024] env[70013]: value = "task-4230970" [ 891.926024] env[70013]: _type = "Task" [ 891.926024] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.926024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.926024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.926024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 891.926740] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-040c2f2c-edfd-450d-93e1-f3dba58ffb71 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.935886] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 891.935886] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b006f-ac04-3bf7-c930-c8097418e730" [ 891.935886] env[70013]: _type = "Task" [ 891.935886] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.940674] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230970, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.951338] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b006f-ac04-3bf7-c930-c8097418e730, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.296654] env[70013]: DEBUG nova.network.neutron [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Updating instance_info_cache with network_info: [{"id": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "address": "fa:16:3e:cd:a2:20", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ae3868-dd", "ovs_interfaceid": "25ae3868-dd45-4f4c-8b91-809dcb235f5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.297071] env[70013]: DEBUG nova.network.neutron [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Updated VIF entry in instance network info cache for port b1d1b3a0-6686-4faf-981e-6772c3a821dd. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.297415] env[70013]: DEBUG nova.network.neutron [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Updating instance_info_cache with network_info: [{"id": "b1d1b3a0-6686-4faf-981e-6772c3a821dd", "address": "fa:16:3e:a9:59:92", "network": {"id": "c639398a-1ce5-43bf-937a-a66b622e33f1", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-565920473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9b70a7428848b7b1f0c618fd9edffc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1d1b3a0-66", "ovs_interfaceid": "b1d1b3a0-6686-4faf-981e-6772c3a821dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.401943] env[70013]: DEBUG nova.compute.utils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 892.403349] env[70013]: DEBUG nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 892.403546] env[70013]: DEBUG nova.network.neutron [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 892.417322] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 892.439746] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230970, 'name': CreateVM_Task, 'duration_secs': 0.333067} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.446167] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.451966] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.456190] env[70013]: DEBUG nova.policy [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd357d73286774d2cbe0ab68494c7c57b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75ec7d6e9ea54c06bd3b7605f16b391a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 892.464272] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522b006f-ac04-3bf7-c930-c8097418e730, 'name': SearchDatastore_Task, 'duration_secs': 0.020456} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.468021] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.468344] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.468640] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.468835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 892.469109] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.469760] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 892.470356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 892.470652] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecd6b9e7-a6ea-4bc2-9bed-1a68aa104ea3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.473223] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9de45bd-d4ad-4ef0-837f-e456a0a8951f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.480119] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 892.480119] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]520b71fc-154d-b801-35b0-aab24d47ca6a" [ 892.480119] env[70013]: _type = "Task" [ 892.480119] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.489394] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.489676] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 892.491479] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96dfd953-4189-4ea2-932d-ef4c95012571 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.501979] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520b71fc-154d-b801-35b0-aab24d47ca6a, 'name': SearchDatastore_Task, 'duration_secs': 0.010745} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.507373] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.507373] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.507373] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.510260] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 892.510260] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd60ed-382c-90fd-2dc6-bbaa7b41a19a" [ 892.510260] env[70013]: _type = "Task" [ 892.510260] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.521470] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd60ed-382c-90fd-2dc6-bbaa7b41a19a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.541614] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327448cb-f6e3-4d0a-af7a-7741da70e64f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.552214] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c93046-bd3a-4b4c-ad56-11f8e1b0e4fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.597623] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5714d9-15b3-4aac-afb1-72cdaa8b7cc6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.606734] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610395f7-55b6-4f24-9ab7-201132c2056c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.622538] env[70013]: DEBUG nova.compute.provider_tree [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.787046] env[70013]: DEBUG nova.network.neutron [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Successfully created port: f777b44c-5cf6-4ccb-86d9-15dad465d158 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.799970] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "refresh_cache-ef64a05a-b514-4c35-81d3-664ae1ad3ff1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.802377] env[70013]: DEBUG oslo_concurrency.lockutils [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] Releasing lock "refresh_cache-0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.802609] env[70013]: DEBUG nova.compute.manager [req-19a70b09-220b-4d77-a955-416f4965e334 req-ec08fdc2-f3b8-4ab8-8841-864c6380414b service nova] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Received event network-vif-deleted-95450791-e5ab-4650-be74-3d6299da9718 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 892.906771] env[70013]: DEBUG nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 893.023257] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd60ed-382c-90fd-2dc6-bbaa7b41a19a, 'name': SearchDatastore_Task, 'duration_secs': 0.010952} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.024092] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1820c52-e745-466a-8a48-5646ed7a6aa9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.031121] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 893.031121] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fe318b-1d21-ee25-4a13-fb53595a70dc" [ 893.031121] env[70013]: _type = "Task" [ 893.031121] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.039058] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fe318b-1d21-ee25-4a13-fb53595a70dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.125791] env[70013]: DEBUG nova.scheduler.client.report [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.541314] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fe318b-1d21-ee25-4a13-fb53595a70dc, 'name': SearchDatastore_Task, 'duration_secs': 0.012852} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.541504] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 893.542165] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff/9137a6bc-8f1e-47b7-b1ec-5b4035d36fff.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.542165] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 893.542357] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.542572] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1313697b-e0a2-46e8-a4b8-b32a5c85846b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.544740] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-223036f8-e59c-42d4-a815-82e38a5acead {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.552201] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 893.552201] env[70013]: value = "task-4230971" [ 893.552201] env[70013]: _type = "Task" [ 893.552201] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.557378] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.557650] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 893.559136] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf7b7cd0-d5dc-43f7-9231-95f273acc963 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.566288] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230971, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.570479] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 893.570479] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ebf4ad-2f5b-76da-8e17-e4d8f37f5a42" [ 893.570479] env[70013]: _type = "Task" [ 893.570479] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.581942] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ebf4ad-2f5b-76da-8e17-e4d8f37f5a42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.635382] env[70013]: DEBUG oslo_concurrency.lockutils [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.724s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.635770] env[70013]: INFO nova.compute.manager [None req-094de637-3240-4e86-8b18-f58c636b4806 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Successfully reverted task state from rebuilding on failure for instance. [ 893.646481] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.550s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 893.646889] env[70013]: DEBUG nova.objects.instance [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'resources' on Instance uuid f715cf45-81f1-4979-abfc-455b3e64f00e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.933132] env[70013]: DEBUG nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 893.966316] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 893.966613] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.966856] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 893.967381] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.967621] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 893.967799] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 893.968084] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 893.968364] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 893.968621] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 893.968826] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 893.969030] env[70013]: DEBUG nova.virt.hardware [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 893.970131] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0acc31-09db-4a68-bed4-6feb9c45d63b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.981556] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4ce73d-623e-473f-9556-07b1c1df73f2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.062665] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230971, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.081859] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ebf4ad-2f5b-76da-8e17-e4d8f37f5a42, 'name': SearchDatastore_Task, 'duration_secs': 0.010017} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.082952] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0268c7ee-813a-4ebd-9727-e967df383d3a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.091960] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 894.091960] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]520f629e-50c7-6003-6e79-6d53ec945115" [ 894.091960] env[70013]: _type = "Task" [ 894.091960] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.102197] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520f629e-50c7-6003-6e79-6d53ec945115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.309528] env[70013]: DEBUG nova.compute.manager [req-b29b1462-6258-4fd0-bade-aabc2f3cd875 req-0ba12e3a-dba4-441e-b5b7-1cacceb0800a service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Received event network-vif-plugged-f777b44c-5cf6-4ccb-86d9-15dad465d158 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 894.309770] env[70013]: DEBUG oslo_concurrency.lockutils [req-b29b1462-6258-4fd0-bade-aabc2f3cd875 req-0ba12e3a-dba4-441e-b5b7-1cacceb0800a service nova] Acquiring lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.310022] env[70013]: DEBUG oslo_concurrency.lockutils [req-b29b1462-6258-4fd0-bade-aabc2f3cd875 req-0ba12e3a-dba4-441e-b5b7-1cacceb0800a service nova] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.310208] env[70013]: DEBUG oslo_concurrency.lockutils [req-b29b1462-6258-4fd0-bade-aabc2f3cd875 req-0ba12e3a-dba4-441e-b5b7-1cacceb0800a service nova] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.310379] env[70013]: DEBUG nova.compute.manager [req-b29b1462-6258-4fd0-bade-aabc2f3cd875 req-0ba12e3a-dba4-441e-b5b7-1cacceb0800a service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] No waiting events found dispatching network-vif-plugged-f777b44c-5cf6-4ccb-86d9-15dad465d158 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 894.310541] env[70013]: WARNING nova.compute.manager [req-b29b1462-6258-4fd0-bade-aabc2f3cd875 req-0ba12e3a-dba4-441e-b5b7-1cacceb0800a service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Received unexpected event network-vif-plugged-f777b44c-5cf6-4ccb-86d9-15dad465d158 for instance with vm_state building and task_state spawning. [ 894.335406] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.335693] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f853b52e-15b9-45fd-96ca-e094b76944d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.348556] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 894.348556] env[70013]: value = "task-4230972" [ 894.348556] env[70013]: _type = "Task" [ 894.348556] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.365152] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230972, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.366476] env[70013]: DEBUG nova.network.neutron [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Successfully updated port: f777b44c-5cf6-4ccb-86d9-15dad465d158 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.569162] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230971, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518577} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.569467] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff/9137a6bc-8f1e-47b7-b1ec-5b4035d36fff.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 894.569692] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.569978] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-91587dc8-4ae2-41b5-a6c1-3436aa5ba6bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.582493] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 894.582493] env[70013]: value = "task-4230973" [ 894.582493] env[70013]: _type = "Task" [ 894.582493] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.596433] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.610042] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520f629e-50c7-6003-6e79-6d53ec945115, 'name': SearchDatastore_Task, 'duration_secs': 0.013785} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.610199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.610816] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f/0a56e30f-fb07-4a5e-9d69-9603fd38ff9f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 894.610816] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e848d81-e94c-40db-898e-abd937428846 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.619417] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 894.619417] env[70013]: value = "task-4230974" [ 894.619417] env[70013]: _type = "Task" [ 894.619417] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.628670] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230974, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.720554] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2b5316-7aff-4df1-af08-7e134b841c64 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.730536] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e027d611-2ebf-49ae-9e24-5bfa6da098d0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.774801] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310f236f-c3ef-473c-a85a-242c843a91e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.783545] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb36d0f9-46fa-4b93-bea8-31e2bae6555e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.801470] env[70013]: DEBUG nova.compute.provider_tree [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.860378] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230972, 'name': PowerOffVM_Task, 'duration_secs': 0.220901} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.860723] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.861872] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568a7c8c-a827-43a9-ac21-bc4c27b07c40 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.883171] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.883314] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.884165] env[70013]: DEBUG nova.network.neutron [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.887088] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8683500e-a1ab-42f9-b1b7-150e1fb8a381 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.932767] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.935772] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d363ab9-f6b4-4c2f-9901-27d103ca5745 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.946394] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 894.946394] env[70013]: value = "task-4230975" [ 894.946394] env[70013]: _type = "Task" [ 894.946394] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.964701] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 894.964701] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.964701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.964701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.964877] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.965265] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35df5d8b-44a3-4f72-bdda-bcb8e330e4a0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.986259] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.986481] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 894.987311] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fdd63f8-f3b1-44ef-9997-813e33d5f947 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.994611] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 894.994611] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f5e0b5-989f-49e9-61b3-c1725ef8d983" [ 894.994611] env[70013]: _type = "Task" [ 894.994611] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.006021] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f5e0b5-989f-49e9-61b3-c1725ef8d983, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.096046] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075562} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.096421] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.097661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcb4164-1dac-4d02-a2da-1ca4351ae325 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.123228] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff/9137a6bc-8f1e-47b7-b1ec-5b4035d36fff.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.123641] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08c61951-74ec-495b-9750-2471ada3d6c7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.148374] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230974, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.150680] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 895.150680] env[70013]: value = "task-4230976" [ 895.150680] env[70013]: _type = "Task" [ 895.150680] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.160757] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.304799] env[70013]: DEBUG nova.scheduler.client.report [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 895.418664] env[70013]: DEBUG nova.network.neutron [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.509311] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f5e0b5-989f-49e9-61b3-c1725ef8d983, 'name': SearchDatastore_Task, 'duration_secs': 0.052384} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.509311] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad2c9f69-9509-48ad-a56f-09a87f97aa93 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.516151] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 895.516151] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52179aa7-a7b6-9b64-eeed-e19f000f5df5" [ 895.516151] env[70013]: _type = "Task" [ 895.516151] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.525196] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52179aa7-a7b6-9b64-eeed-e19f000f5df5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.633963] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230974, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605632} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.634287] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f/0a56e30f-fb07-4a5e-9d69-9603fd38ff9f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 895.634506] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.634763] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b48c04c8-51bb-47f6-b4b4-c0c05b5d2bfa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.637799] env[70013]: DEBUG nova.network.neutron [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Updating instance_info_cache with network_info: [{"id": "f777b44c-5cf6-4ccb-86d9-15dad465d158", "address": "fa:16:3e:ef:4d:91", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf777b44c-5c", "ovs_interfaceid": "f777b44c-5cf6-4ccb-86d9-15dad465d158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.646449] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 895.646449] env[70013]: value = "task-4230977" [ 895.646449] env[70013]: _type = "Task" [ 895.646449] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.663800] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.668110] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230976, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.811189] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.165s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 895.814738] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.305s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.815831] env[70013]: INFO nova.compute.claims [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.841702] env[70013]: INFO nova.scheduler.client.report [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted allocations for instance f715cf45-81f1-4979-abfc-455b3e64f00e [ 896.027222] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52179aa7-a7b6-9b64-eeed-e19f000f5df5, 'name': SearchDatastore_Task, 'duration_secs': 0.019095} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.027549] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 896.027878] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. {{(pid=70013) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 896.028226] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8531bdb0-ba22-41f2-9100-6ff88cf0fc9f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.036753] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 896.036753] env[70013]: value = "task-4230978" [ 896.036753] env[70013]: _type = "Task" [ 896.036753] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.046594] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230978, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.141135] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 896.142194] env[70013]: DEBUG nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance network_info: |[{"id": "f777b44c-5cf6-4ccb-86d9-15dad465d158", "address": "fa:16:3e:ef:4d:91", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf777b44c-5c", "ovs_interfaceid": "f777b44c-5cf6-4ccb-86d9-15dad465d158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 896.142194] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:4d:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f777b44c-5cf6-4ccb-86d9-15dad465d158', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.150167] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 896.152050] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.153924] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-058b2cbd-ea94-4849-a283-654fd5acfb2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.177879] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230976, 'name': ReconfigVM_Task, 'duration_secs': 0.601041} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.182023] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff/9137a6bc-8f1e-47b7-b1ec-5b4035d36fff.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.182727] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.182727] env[70013]: value = "task-4230979" [ 896.182727] env[70013]: _type = "Task" [ 896.182727] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.182973] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.249206} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.183228] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14682df1-8a6e-45bd-ad91-14532b239f48 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.185080] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.188502] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b5a54b-1115-459f-acef-7f21bb91b136 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.198570] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230979, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.218471] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f/0a56e30f-fb07-4a5e-9d69-9603fd38ff9f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.218942] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 896.218942] env[70013]: value = "task-4230980" [ 896.218942] env[70013]: _type = "Task" [ 896.218942] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.219173] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3f3c9d9-9c86-4a32-826d-19cb88b7ea36 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.244756] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230980, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.246486] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 896.246486] env[70013]: value = "task-4230981" [ 896.246486] env[70013]: _type = "Task" [ 896.246486] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.256376] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230981, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.349697] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b1dcdbfa-4126-4acb-b76e-4b87e0f7ede0 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "f715cf45-81f1-4979-abfc-455b3e64f00e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.808s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.414508] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 896.414774] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.414958] env[70013]: DEBUG nova.compute.manager [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 896.415914] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a2d5eb-a617-4f46-8716-94b9b49e2012 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.424181] env[70013]: DEBUG nova.compute.manager [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 896.425047] env[70013]: DEBUG nova.objects.instance [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'flavor' on Instance uuid 28a56a75-9b30-4121-8252-a9e57287441c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.476667] env[70013]: DEBUG nova.compute.manager [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Received event network-changed-f777b44c-5cf6-4ccb-86d9-15dad465d158 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 896.476957] env[70013]: DEBUG nova.compute.manager [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Refreshing instance network info cache due to event network-changed-f777b44c-5cf6-4ccb-86d9-15dad465d158. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 896.477310] env[70013]: DEBUG oslo_concurrency.lockutils [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] Acquiring lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.478414] env[70013]: DEBUG oslo_concurrency.lockutils [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] Acquired lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 896.478774] env[70013]: DEBUG nova.network.neutron [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Refreshing network info cache for port f777b44c-5cf6-4ccb-86d9-15dad465d158 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.549848] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230978, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.696382] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4230979, 'name': CreateVM_Task, 'duration_secs': 0.445297} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.696580] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.697419] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.697526] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 896.697954] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 896.698331] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08e4b1ca-8306-4692-a19b-cc13f94845c3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.707878] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 896.707878] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5294b31a-595c-9b0b-414a-6b9f8a831514" [ 896.707878] env[70013]: _type = "Task" [ 896.707878] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.722376] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5294b31a-595c-9b0b-414a-6b9f8a831514, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.747882] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230980, 'name': Rename_Task, 'duration_secs': 0.209808} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.752333] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.752667] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09817a6b-1961-44e0-9675-5d2ccc2e7fe6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.762115] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230981, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.763933] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 896.763933] env[70013]: value = "task-4230982" [ 896.763933] env[70013]: _type = "Task" [ 896.763933] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.777617] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.051504] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230978, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.834961} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.051504] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. [ 897.051504] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441f94c3-86d9-4d63-a098-fa3aa599aa16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.088043] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.094468] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2afd942a-f554-4ff7-b81a-5cbbb27ec3ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.113898] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 897.113898] env[70013]: value = "task-4230983" [ 897.113898] env[70013]: _type = "Task" [ 897.113898] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.124289] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230983, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.226168] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5294b31a-595c-9b0b-414a-6b9f8a831514, 'name': SearchDatastore_Task, 'duration_secs': 0.068197} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.227316] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 897.227316] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.227316] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.227498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 897.227619] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.227972] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a3c35f3-a182-4bfb-bea4-2cb05fbde624 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.244395] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.244602] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 897.245420] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd1d1341-ffa5-412a-8ea1-2815b3561445 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.264232] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 897.264232] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5219ec81-9f4c-6199-1236-98d4e9e4db97" [ 897.264232] env[70013]: _type = "Task" [ 897.264232] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.268257] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230981, 'name': ReconfigVM_Task, 'duration_secs': 0.865575} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.280849] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f/0a56e30f-fb07-4a5e-9d69-9603fd38ff9f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.284859] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-317f9286-9eb5-4a61-9661-6944611ea8bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.295536] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5219ec81-9f4c-6199-1236-98d4e9e4db97, 'name': SearchDatastore_Task, 'duration_secs': 0.018254} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.302708] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 897.302708] env[70013]: value = "task-4230984" [ 897.302708] env[70013]: _type = "Task" [ 897.302708] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.302998] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230982, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.303346] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29b38950-6253-4bf1-9b49-da944f801b11 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.319171] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 897.319171] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529fcbc0-5aee-7c82-02b0-6db382315219" [ 897.319171] env[70013]: _type = "Task" [ 897.319171] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.323775] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230984, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.339479] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529fcbc0-5aee-7c82-02b0-6db382315219, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.345183] env[70013]: DEBUG nova.network.neutron [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Updated VIF entry in instance network info cache for port f777b44c-5cf6-4ccb-86d9-15dad465d158. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.346028] env[70013]: DEBUG nova.network.neutron [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Updating instance_info_cache with network_info: [{"id": "f777b44c-5cf6-4ccb-86d9-15dad465d158", "address": "fa:16:3e:ef:4d:91", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf777b44c-5c", "ovs_interfaceid": "f777b44c-5cf6-4ccb-86d9-15dad465d158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.347212] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "dacc5f9c-3cd2-4992-b59a-260d2657567e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.347448] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "dacc5f9c-3cd2-4992-b59a-260d2657567e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 897.439986] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.440451] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0d14d52-a846-4c4a-8771-53465b15af5d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.448163] env[70013]: DEBUG oslo_vmware.api [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 897.448163] env[70013]: value = "task-4230985" [ 897.448163] env[70013]: _type = "Task" [ 897.448163] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.454124] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35afb58e-819c-4df0-a84f-3372ea62a5ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.468583] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c09eaa3-2476-40bc-af84-d53aace52b0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.475216] env[70013]: DEBUG oslo_vmware.api [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230985, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.508125] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be67630b-ac80-4a18-a2e3-774eea356761 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.518990] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062bd5f4-0de1-467f-8e80-8f990502a0ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.535766] env[70013]: DEBUG nova.compute.provider_tree [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.625742] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230983, 'name': ReconfigVM_Task, 'duration_secs': 0.402818} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.625979] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Reconfigured VM instance instance-00000033 to attach disk [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.626858] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55462dc-ba7a-4329-806e-b314f42e974f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.654595] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-929ffdcb-02c7-4704-b20e-797d58119737 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.672709] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 897.672709] env[70013]: value = "task-4230986" [ 897.672709] env[70013]: _type = "Task" [ 897.672709] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.682935] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230986, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.778347] env[70013]: DEBUG oslo_vmware.api [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4230982, 'name': PowerOnVM_Task, 'duration_secs': 0.56427} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.778672] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.778822] env[70013]: INFO nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Took 11.89 seconds to spawn the instance on the hypervisor. [ 897.779014] env[70013]: DEBUG nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 897.779837] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15e390a-6672-457b-8446-0fdfcee4d1b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.820780] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230984, 'name': Rename_Task, 'duration_secs': 0.400003} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.821035] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.821308] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44a21211-4307-4478-961a-79ce93601060 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.829766] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 897.829766] env[70013]: value = "task-4230987" [ 897.829766] env[70013]: _type = "Task" [ 897.829766] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.833815] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529fcbc0-5aee-7c82-02b0-6db382315219, 'name': SearchDatastore_Task, 'duration_secs': 0.029917} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.837337] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 897.837337] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] eab0c393-4d3d-4659-8225-c82ddadb330b/eab0c393-4d3d-4659-8225-c82ddadb330b.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.837563] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84df3fef-dfb0-4892-b506-e8ab0af0a582 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.846994] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230987, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.848971] env[70013]: DEBUG oslo_concurrency.lockutils [req-730eddea-8ce6-419c-90ca-75119d3ef671 req-d2598f95-aa42-4b2d-807a-3f19782dad9f service nova] Releasing lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 897.849424] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 897.849424] env[70013]: value = "task-4230988" [ 897.849424] env[70013]: _type = "Task" [ 897.849424] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.859243] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.959065] env[70013]: DEBUG oslo_vmware.api [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230985, 'name': PowerOffVM_Task, 'duration_secs': 0.431473} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.959365] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.959565] env[70013]: DEBUG nova.compute.manager [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 897.960589] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5fcf47-0c71-4ab0-8fc4-5652774f174a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.038903] env[70013]: DEBUG nova.scheduler.client.report [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 898.188204] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230986, 'name': ReconfigVM_Task, 'duration_secs': 0.246382} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.188659] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.189064] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e234c3f9-d690-4bff-8552-a491f1a0d2b2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.198721] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 898.198721] env[70013]: value = "task-4230989" [ 898.198721] env[70013]: _type = "Task" [ 898.198721] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.211857] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230989, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.304466] env[70013]: INFO nova.compute.manager [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Took 39.25 seconds to build instance. [ 898.345567] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230987, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.362337] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230988, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.477356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0eb2a29a-8644-4c0b-ad97-1ffc62dc8891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.062s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.550677] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.736s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.551319] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 898.556132] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.106s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.558563] env[70013]: INFO nova.compute.claims [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.710958] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230989, 'name': PowerOnVM_Task} progress is 90%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.807744] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748ffec7-cc23-4672-abba-564fc5ea686c tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.060s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.846912] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230987, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.862183] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705582} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.862577] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] eab0c393-4d3d-4659-8225-c82ddadb330b/eab0c393-4d3d-4659-8225-c82ddadb330b.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.862892] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.863079] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b568aa2a-2932-47c9-a0b7-806d3a4a0a51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.871925] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 898.871925] env[70013]: value = "task-4230990" [ 898.871925] env[70013]: _type = "Task" [ 898.871925] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.884432] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230990, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.057303] env[70013]: DEBUG nova.compute.utils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 899.058849] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 899.059471] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.113215] env[70013]: DEBUG nova.policy [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cc65d5bfee34b9194167bafd0f33e0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4c67d4c6cfb40d08ae8ee115ade7273', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 899.210045] env[70013]: DEBUG oslo_vmware.api [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4230989, 'name': PowerOnVM_Task, 'duration_secs': 0.731504} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.210387] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.214953] env[70013]: DEBUG nova.compute.manager [None req-f0ad47c1-237a-4b1a-b1c5-fb3dfb4fe520 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 899.215905] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718047ca-b2b3-44d1-9915-90988774f243 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.283148] env[70013]: DEBUG nova.objects.instance [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'flavor' on Instance uuid 28a56a75-9b30-4121-8252-a9e57287441c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.311849] env[70013]: DEBUG nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 899.347021] env[70013]: DEBUG oslo_vmware.api [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230987, 'name': PowerOnVM_Task, 'duration_secs': 1.041817} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.348477] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.348990] env[70013]: INFO nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Took 10.83 seconds to spawn the instance on the hypervisor. [ 899.348990] env[70013]: DEBUG nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 899.350442] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4194621-1456-45e7-8823-deef81b7c7c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.386466] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230990, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078027} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.391560] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.392532] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7024308-095d-4f15-921d-24d46f378cb9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.426577] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] eab0c393-4d3d-4659-8225-c82ddadb330b/eab0c393-4d3d-4659-8225-c82ddadb330b.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.427881] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-151ba6fe-2e83-47b8-9d76-b8ca31c537c8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.457113] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 899.457113] env[70013]: value = "task-4230991" [ 899.457113] env[70013]: _type = "Task" [ 899.457113] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.470998] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230991, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.517911] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Successfully created port: c7d197fd-939c-426a-b373-674e428f6655 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.563572] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 899.790345] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.790614] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.790845] env[70013]: DEBUG nova.network.neutron [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.791118] env[70013]: DEBUG nova.objects.instance [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'info_cache' on Instance uuid 28a56a75-9b30-4121-8252-a9e57287441c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.834617] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.888195] env[70013]: INFO nova.compute.manager [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Took 39.78 seconds to build instance. [ 899.920928] env[70013]: DEBUG nova.compute.manager [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Received event network-changed-04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 899.920928] env[70013]: DEBUG nova.compute.manager [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Refreshing instance network info cache due to event network-changed-04205eac-90e1-47c0-82c4-703b9258de68. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 899.920928] env[70013]: DEBUG oslo_concurrency.lockutils [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] Acquiring lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.920928] env[70013]: DEBUG oslo_concurrency.lockutils [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] Acquired lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.920928] env[70013]: DEBUG nova.network.neutron [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Refreshing network info cache for port 04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.969496] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230991, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.080194] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6918a3-500b-4996-b192-67ff85171aec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.089198] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd9cf54-ae92-4e93-b548-5271ad33a63d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.124834] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a66956f-abba-4fc1-bf65-e7178994522c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.135710] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8897bb-719a-4355-ae07-d8b56a1a1cc7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.153093] env[70013]: DEBUG nova.compute.provider_tree [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.294948] env[70013]: DEBUG nova.objects.base [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Object Instance<28a56a75-9b30-4121-8252-a9e57287441c> lazy-loaded attributes: flavor,info_cache {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 900.395346] env[70013]: DEBUG oslo_concurrency.lockutils [None req-23956bd6-89e3-4399-aeda-f36f747f2f5b tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.271s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.471161] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230991, 'name': ReconfigVM_Task, 'duration_secs': 0.613501} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.474270] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Reconfigured VM instance instance-00000036 to attach disk [datastore1] eab0c393-4d3d-4659-8225-c82ddadb330b/eab0c393-4d3d-4659-8225-c82ddadb330b.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.475583] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ab4cebe-869f-4958-9caf-e49421ecf631 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.485200] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 900.485200] env[70013]: value = "task-4230992" [ 900.485200] env[70013]: _type = "Task" [ 900.485200] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.495785] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230992, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.574555] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 900.599899] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 900.600119] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.600291] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 900.600473] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.600624] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 900.600769] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 900.600998] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 900.601192] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 900.601360] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 900.601518] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 900.601687] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 900.602679] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2225ece-c29a-4ad1-8f60-01032fe3e561 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.613530] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62247a81-6dfe-440e-933e-b7e280fe17bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.658922] env[70013]: DEBUG nova.scheduler.client.report [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 900.717483] env[70013]: DEBUG nova.network.neutron [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Updated VIF entry in instance network info cache for port 04205eac-90e1-47c0-82c4-703b9258de68. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.717972] env[70013]: DEBUG nova.network.neutron [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Updating instance_info_cache with network_info: [{"id": "04205eac-90e1-47c0-82c4-703b9258de68", "address": "fa:16:3e:8e:33:9e", "network": {"id": "1d6a6584-952d-4f79-ba84-877f008c3ba9", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-840362347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f16290bcd228490fbdb41439404d61d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04205eac-90", "ovs_interfaceid": "04205eac-90e1-47c0-82c4-703b9258de68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.900686] env[70013]: DEBUG nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 901.000726] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230992, 'name': Rename_Task, 'duration_secs': 0.370573} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.001182] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.001594] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d4beca4-8e3c-49b0-870f-887a03504631 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.012368] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 901.012368] env[70013]: value = "task-4230993" [ 901.012368] env[70013]: _type = "Task" [ 901.012368] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.026224] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.165372] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.165372] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 901.171445] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.928s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.171445] env[70013]: DEBUG nova.objects.instance [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lazy-loading 'resources' on Instance uuid 5c5fbdae-4916-4beb-ada0-57f36fb0f84c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.174633] env[70013]: DEBUG nova.compute.manager [req-5400581b-ea57-42fa-995c-941525b07288 req-7f0add42-a16a-40cb-b010-607aaccc1f86 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Received event network-vif-plugged-c7d197fd-939c-426a-b373-674e428f6655 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 901.175040] env[70013]: DEBUG oslo_concurrency.lockutils [req-5400581b-ea57-42fa-995c-941525b07288 req-7f0add42-a16a-40cb-b010-607aaccc1f86 service nova] Acquiring lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.175180] env[70013]: DEBUG oslo_concurrency.lockutils [req-5400581b-ea57-42fa-995c-941525b07288 req-7f0add42-a16a-40cb-b010-607aaccc1f86 service nova] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.175391] env[70013]: DEBUG oslo_concurrency.lockutils [req-5400581b-ea57-42fa-995c-941525b07288 req-7f0add42-a16a-40cb-b010-607aaccc1f86 service nova] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.175599] env[70013]: DEBUG nova.compute.manager [req-5400581b-ea57-42fa-995c-941525b07288 req-7f0add42-a16a-40cb-b010-607aaccc1f86 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] No waiting events found dispatching network-vif-plugged-c7d197fd-939c-426a-b373-674e428f6655 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 901.175787] env[70013]: WARNING nova.compute.manager [req-5400581b-ea57-42fa-995c-941525b07288 req-7f0add42-a16a-40cb-b010-607aaccc1f86 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Received unexpected event network-vif-plugged-c7d197fd-939c-426a-b373-674e428f6655 for instance with vm_state building and task_state spawning. [ 901.215757] env[70013]: DEBUG nova.network.neutron [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.221404] env[70013]: DEBUG oslo_concurrency.lockutils [req-46953516-9548-4725-9044-678b2be5617b req-307371ac-4837-49db-8de4-484f95de2815 service nova] Releasing lock "refresh_cache-9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.243854] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.243854] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.244103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.244103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.244286] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.249504] env[70013]: INFO nova.compute.manager [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Terminating instance [ 901.279349] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Successfully updated port: c7d197fd-939c-426a-b373-674e428f6655 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.426873] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.524859] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230993, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.675435] env[70013]: DEBUG nova.compute.utils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 901.682189] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 901.682372] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 901.718680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.754636] env[70013]: DEBUG nova.compute.manager [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 901.754636] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.755518] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f375c605-2039-446b-b405-13e923d519bd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.766199] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.766758] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44b28f92-4462-4b3b-b2d9-e7a02056bae8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.772052] env[70013]: DEBUG nova.policy [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cc65d5bfee34b9194167bafd0f33e0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4c67d4c6cfb40d08ae8ee115ade7273', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 901.775592] env[70013]: DEBUG oslo_vmware.api [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 901.775592] env[70013]: value = "task-4230994" [ 901.775592] env[70013]: _type = "Task" [ 901.775592] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.791352] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "refresh_cache-d84d8426-b3fd-4f0d-8410-0ee24fedb35d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.791544] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "refresh_cache-d84d8426-b3fd-4f0d-8410-0ee24fedb35d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.791859] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.797227] env[70013]: DEBUG oslo_vmware.api [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.023854] env[70013]: DEBUG oslo_vmware.api [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4230993, 'name': PowerOnVM_Task, 'duration_secs': 0.890165} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.023997] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.024420] env[70013]: INFO nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Took 8.09 seconds to spawn the instance on the hypervisor. [ 902.024420] env[70013]: DEBUG nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 902.026165] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed27780-43c1-46c7-8571-35c8546ce9ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.128379] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Successfully created port: 7825762a-5525-4c92-8146-1ac33ba295f6 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.184823] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 902.260148] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa21efa-181d-439c-9561-485da2d853ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.268873] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6703edcf-0e5f-4bbd-b9f3-4cbf4ee3f3ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.315758] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7457b76-82ea-4546-b046-500e3183d7fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.327396] env[70013]: DEBUG oslo_vmware.api [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230994, 'name': PowerOffVM_Task, 'duration_secs': 0.258877} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.328299] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a0c422-b78f-4dfa-ad63-861b66ce2dd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.332681] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.332865] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.334086] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8149596-99cf-4be7-9efc-ac8d7e178b3c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.351782] env[70013]: DEBUG nova.compute.provider_tree [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.368175] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 902.540789] env[70013]: INFO nova.compute.manager [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Rescuing [ 902.541117] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.541271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.541441] env[70013]: DEBUG nova.network.neutron [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.569905] env[70013]: INFO nova.compute.manager [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Took 41.51 seconds to build instance. [ 902.655951] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Updating instance_info_cache with network_info: [{"id": "c7d197fd-939c-426a-b373-674e428f6655", "address": "fa:16:3e:f0:e5:0c", "network": {"id": "f0029e8a-cbc6-4c49-b026-38d3aa38ee1c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-216013226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4c67d4c6cfb40d08ae8ee115ade7273", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7d197fd-93", "ovs_interfaceid": "c7d197fd-939c-426a-b373-674e428f6655", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.728032] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.728371] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00b45db7-bdd0-43bf-a62b-99e1d87d04d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.738181] env[70013]: DEBUG oslo_vmware.api [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 902.738181] env[70013]: value = "task-4230996" [ 902.738181] env[70013]: _type = "Task" [ 902.738181] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.747297] env[70013]: DEBUG oslo_vmware.api [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.859029] env[70013]: DEBUG nova.scheduler.client.report [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.071965] env[70013]: DEBUG oslo_concurrency.lockutils [None req-456faabb-7d57-41f5-ba8b-f03aba773d54 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.752s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.163380] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "refresh_cache-d84d8426-b3fd-4f0d-8410-0ee24fedb35d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.163739] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Instance network_info: |[{"id": "c7d197fd-939c-426a-b373-674e428f6655", "address": "fa:16:3e:f0:e5:0c", "network": {"id": "f0029e8a-cbc6-4c49-b026-38d3aa38ee1c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-216013226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4c67d4c6cfb40d08ae8ee115ade7273", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7d197fd-93", "ovs_interfaceid": "c7d197fd-939c-426a-b373-674e428f6655", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 903.164675] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:e5:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7d197fd-939c-426a-b373-674e428f6655', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.173463] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Creating folder: Project (b4c67d4c6cfb40d08ae8ee115ade7273). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 903.175238] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae62c858-ce79-4342-a083-2b183f40886b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.177033] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.177737] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.177935] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Deleting the datastore file [datastore1] 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.178218] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-115506f2-6496-4d89-8d60-4f7b549f9199 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.186290] env[70013]: DEBUG oslo_vmware.api [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for the task: (returnval){ [ 903.186290] env[70013]: value = "task-4230997" [ 903.186290] env[70013]: _type = "Task" [ 903.186290] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.191347] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Created folder: Project (b4c67d4c6cfb40d08ae8ee115ade7273) in parent group-v836999. [ 903.191549] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Creating folder: Instances. Parent ref: group-v837134. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 903.192438] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e2c48ad9-9772-41f5-8a68-1143d45aba91 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.197509] env[70013]: DEBUG oslo_vmware.api [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230997, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.203065] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 903.206482] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Created folder: Instances in parent group-v837134. [ 903.206482] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 903.206600] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.206802] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37abcb07-48b5-423c-9578-d273fe4d64aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.229195] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.229195] env[70013]: value = "task-4231000" [ 903.229195] env[70013]: _type = "Task" [ 903.229195] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.241952] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 903.242511] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.242633] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 903.242828] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.243098] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 903.243295] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 903.243537] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 903.243696] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 903.243948] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 903.244297] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 903.244297] env[70013]: DEBUG nova.virt.hardware [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 903.245584] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c084d1c-d25b-406c-be9a-767b002eb222 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.258632] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231000, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.266732] env[70013]: DEBUG oslo_vmware.api [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230996, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.268339] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1346bd1-807a-49c3-ae3a-d9017d7b2223 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.365649] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.196s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.368066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.545s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.368066] env[70013]: DEBUG nova.objects.instance [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lazy-loading 'resources' on Instance uuid b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.451922] env[70013]: DEBUG nova.compute.manager [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Received event network-changed-c7d197fd-939c-426a-b373-674e428f6655 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 903.452180] env[70013]: DEBUG nova.compute.manager [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Refreshing instance network info cache due to event network-changed-c7d197fd-939c-426a-b373-674e428f6655. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 903.452902] env[70013]: DEBUG oslo_concurrency.lockutils [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] Acquiring lock "refresh_cache-d84d8426-b3fd-4f0d-8410-0ee24fedb35d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.452902] env[70013]: DEBUG oslo_concurrency.lockutils [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] Acquired lock "refresh_cache-d84d8426-b3fd-4f0d-8410-0ee24fedb35d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.452902] env[70013]: DEBUG nova.network.neutron [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Refreshing network info cache for port c7d197fd-939c-426a-b373-674e428f6655 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 903.461433] env[70013]: DEBUG nova.network.neutron [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Updating instance_info_cache with network_info: [{"id": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "address": "fa:16:3e:01:6f:a2", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a9242b0-4a", "ovs_interfaceid": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.575096] env[70013]: DEBUG nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 903.641420] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 903.641420] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 903.699228] env[70013]: DEBUG oslo_vmware.api [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Task: {'id': task-4230997, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165562} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.699586] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.699813] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.700041] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.700218] env[70013]: INFO nova.compute.manager [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Took 1.95 seconds to destroy the instance on the hypervisor. [ 903.703627] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 903.704452] env[70013]: DEBUG nova.compute.manager [-] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 903.704658] env[70013]: DEBUG nova.network.neutron [-] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 903.744434] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231000, 'name': CreateVM_Task, 'duration_secs': 0.399275} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.744726] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 903.748434] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.748611] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.748904] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 903.749225] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99dfe07a-54e8-4f63-b86f-cd15afba24d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.755740] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 903.755740] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5209220b-3297-c58f-fe88-78e10b4eb271" [ 903.755740] env[70013]: _type = "Task" [ 903.755740] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.761100] env[70013]: DEBUG oslo_vmware.api [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230996, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.772855] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5209220b-3297-c58f-fe88-78e10b4eb271, 'name': SearchDatastore_Task, 'duration_secs': 0.010278} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.773184] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.773429] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.773665] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.773808] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.774074] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.774260] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-395ad630-7ba7-4c8d-8458-cd79ab7f519a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.784581] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.784700] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 903.785766] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bfe44a5-f801-4809-8ade-024c56aca2a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.794113] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 903.794113] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52121946-d0a5-a2ae-008d-c8d2ecab7b51" [ 903.794113] env[70013]: _type = "Task" [ 903.794113] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.806599] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52121946-d0a5-a2ae-008d-c8d2ecab7b51, 'name': SearchDatastore_Task, 'duration_secs': 0.010186} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.807490] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06c58340-154e-4af1-bf79-94d1f712bf18 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.813503] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 903.813503] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ee1e95-0452-33ce-38de-195ba9b0bd35" [ 903.813503] env[70013]: _type = "Task" [ 903.813503] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.823525] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ee1e95-0452-33ce-38de-195ba9b0bd35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.899263] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ac79e0c-6557-4fed-83ca-3c5994e02748 tempest-ServerActionsV293TestJSON-630636606 tempest-ServerActionsV293TestJSON-630636606-project-member] Lock "5c5fbdae-4916-4beb-ada0-57f36fb0f84c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.690s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.970284] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 904.108052] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.154553] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.154553] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.154553] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.154553] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.154553] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.154553] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.154553] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 904.154553] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.260419] env[70013]: DEBUG oslo_vmware.api [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230996, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.329976] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ee1e95-0452-33ce-38de-195ba9b0bd35, 'name': SearchDatastore_Task, 'duration_secs': 0.017645} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.330361] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 904.330588] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] d84d8426-b3fd-4f0d-8410-0ee24fedb35d/d84d8426-b3fd-4f0d-8410-0ee24fedb35d.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 904.332430] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a0dded5-e117-4aec-b4c1-7084765403ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.343505] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 904.343505] env[70013]: value = "task-4231001" [ 904.343505] env[70013]: _type = "Task" [ 904.343505] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.354933] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231001, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.376459] env[70013]: DEBUG nova.network.neutron [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Updated VIF entry in instance network info cache for port c7d197fd-939c-426a-b373-674e428f6655. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.377140] env[70013]: DEBUG nova.network.neutron [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Updating instance_info_cache with network_info: [{"id": "c7d197fd-939c-426a-b373-674e428f6655", "address": "fa:16:3e:f0:e5:0c", "network": {"id": "f0029e8a-cbc6-4c49-b026-38d3aa38ee1c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-216013226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4c67d4c6cfb40d08ae8ee115ade7273", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7d197fd-93", "ovs_interfaceid": "c7d197fd-939c-426a-b373-674e428f6655", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.414611] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8114313-f494-47f1-bdd0-fedb7d6f5c71 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.425454] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf33dc1-6664-4526-922c-9eb05302cf1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.463656] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198dc19b-e2a9-43fc-85c3-89f9a67cd29c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.475484] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f70784-03ee-4af1-902d-52404d018dd1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.481986] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Successfully updated port: 7825762a-5525-4c92-8146-1ac33ba295f6 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.497282] env[70013]: DEBUG nova.compute.provider_tree [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.657533] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.714461] env[70013]: DEBUG nova.compute.manager [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 904.715847] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04599420-cf5b-4672-aad4-57f2c50ded44 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.761686] env[70013]: DEBUG oslo_vmware.api [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4230996, 'name': PowerOnVM_Task, 'duration_secs': 2.0002} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.761686] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.761686] env[70013]: DEBUG nova.compute.manager [None req-3dda1798-99ee-4e87-bd51-66a5bc641c35 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 904.762693] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a9eb8c-e2d9-48ba-8a86-2da72ea0ff4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.855993] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231001, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486133} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.856328] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] d84d8426-b3fd-4f0d-8410-0ee24fedb35d/d84d8426-b3fd-4f0d-8410-0ee24fedb35d.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.856547] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.856798] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72a741ea-c070-419c-ae3b-c2c5ecfa5ae2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.866834] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 904.866834] env[70013]: value = "task-4231002" [ 904.866834] env[70013]: _type = "Task" [ 904.866834] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.878660] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.880546] env[70013]: DEBUG oslo_concurrency.lockutils [req-b4bb9574-40f5-4083-8199-0142ca689322 req-943068e4-b040-4e6a-922b-d36f69718460 service nova] Releasing lock "refresh_cache-d84d8426-b3fd-4f0d-8410-0ee24fedb35d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 904.986994] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "refresh_cache-2aefaa91-c439-486a-8b19-c6f45f52583f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.986994] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "refresh_cache-2aefaa91-c439-486a-8b19-c6f45f52583f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.986994] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.002021] env[70013]: DEBUG nova.scheduler.client.report [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.009616] env[70013]: DEBUG nova.network.neutron [-] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.232853] env[70013]: INFO nova.compute.manager [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] instance snapshotting [ 905.236398] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a4f85b-dfea-4525-837d-e6f3bd40a041 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.257707] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc72359-4523-4234-994d-138d8e75087e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.377454] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074673} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.377737] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.378622] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d02f042-6d47-488c-adf9-2f9df420c46c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.401959] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] d84d8426-b3fd-4f0d-8410-0ee24fedb35d/d84d8426-b3fd-4f0d-8410-0ee24fedb35d.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.402363] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f01f14b-89cd-4821-9310-f87c2ed44fcf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.425572] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 905.425572] env[70013]: value = "task-4231003" [ 905.425572] env[70013]: _type = "Task" [ 905.425572] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.436988] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231003, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.509137] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.142s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.513089] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.369s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.513480] env[70013]: DEBUG nova.objects.instance [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lazy-loading 'resources' on Instance uuid be9a565b-c74c-442e-8087-cc4579b1c8e7 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.517056] env[70013]: INFO nova.compute.manager [-] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Took 1.81 seconds to deallocate network for instance. [ 905.517056] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.517820] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebec9670-ee48-4e6c-9a56-5c8d5f38f2a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.529861] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 905.529861] env[70013]: value = "task-4231004" [ 905.529861] env[70013]: _type = "Task" [ 905.529861] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.535253] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 905.545159] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231004, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.553582] env[70013]: INFO nova.scheduler.client.report [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Deleted allocations for instance b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25 [ 905.716390] env[70013]: DEBUG nova.network.neutron [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Updating instance_info_cache with network_info: [{"id": "7825762a-5525-4c92-8146-1ac33ba295f6", "address": "fa:16:3e:17:35:f5", "network": {"id": "f0029e8a-cbc6-4c49-b026-38d3aa38ee1c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-216013226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4c67d4c6cfb40d08ae8ee115ade7273", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7825762a-55", "ovs_interfaceid": "7825762a-5525-4c92-8146-1ac33ba295f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.766409] env[70013]: DEBUG nova.compute.manager [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Received event network-vif-plugged-7825762a-5525-4c92-8146-1ac33ba295f6 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 905.766621] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] Acquiring lock "2aefaa91-c439-486a-8b19-c6f45f52583f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.766748] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.766861] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.767110] env[70013]: DEBUG nova.compute.manager [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] No waiting events found dispatching network-vif-plugged-7825762a-5525-4c92-8146-1ac33ba295f6 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 905.767245] env[70013]: WARNING nova.compute.manager [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Received unexpected event network-vif-plugged-7825762a-5525-4c92-8146-1ac33ba295f6 for instance with vm_state building and task_state spawning. [ 905.767425] env[70013]: DEBUG nova.compute.manager [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Received event network-changed-7825762a-5525-4c92-8146-1ac33ba295f6 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 905.767584] env[70013]: DEBUG nova.compute.manager [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Refreshing instance network info cache due to event network-changed-7825762a-5525-4c92-8146-1ac33ba295f6. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 905.767758] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] Acquiring lock "refresh_cache-2aefaa91-c439-486a-8b19-c6f45f52583f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.771664] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.772061] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4bc0335c-3b28-44e6-a941-d0b7070734d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.780178] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 905.780178] env[70013]: value = "task-4231005" [ 905.780178] env[70013]: _type = "Task" [ 905.780178] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.790237] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231005, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.941445] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.029277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.045197] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231004, 'name': PowerOffVM_Task, 'duration_secs': 0.399782} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.045197] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.046184] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14448850-be0e-4ede-9db6-83685dfe0c65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.076987] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f3d497-8ff2-47f6-869b-f5c35a5e0eb1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.081797] env[70013]: DEBUG oslo_concurrency.lockutils [None req-cfbaee93-5218-44ac-987e-67d5b715ca78 tempest-VolumesAssistedSnapshotsTest-275775559 tempest-VolumesAssistedSnapshotsTest-275775559-project-member] Lock "b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.452s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.134113] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.134462] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5fb0b27-b5b1-4ca9-9ed5-2433c1cafbdf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.147180] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 906.147180] env[70013]: value = "task-4231006" [ 906.147180] env[70013]: _type = "Task" [ 906.147180] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.162159] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 906.162159] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.162159] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.162159] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.162159] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.162556] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-347d38c2-e42f-47ff-ab05-f15c80d35f1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.173827] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.174069] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.177660] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1bfa92b-bcba-46d7-ac38-8400ff56a021 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.186139] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 906.186139] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a8b0f4-2eba-a6c6-7197-a131e65b82a5" [ 906.186139] env[70013]: _type = "Task" [ 906.186139] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.195648] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a8b0f4-2eba-a6c6-7197-a131e65b82a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.218985] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "refresh_cache-2aefaa91-c439-486a-8b19-c6f45f52583f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.219383] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Instance network_info: |[{"id": "7825762a-5525-4c92-8146-1ac33ba295f6", "address": "fa:16:3e:17:35:f5", "network": {"id": "f0029e8a-cbc6-4c49-b026-38d3aa38ee1c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-216013226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4c67d4c6cfb40d08ae8ee115ade7273", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7825762a-55", "ovs_interfaceid": "7825762a-5525-4c92-8146-1ac33ba295f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 906.220151] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] Acquired lock "refresh_cache-2aefaa91-c439-486a-8b19-c6f45f52583f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.220151] env[70013]: DEBUG nova.network.neutron [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Refreshing network info cache for port 7825762a-5525-4c92-8146-1ac33ba295f6 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.221676] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:35:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7825762a-5525-4c92-8146-1ac33ba295f6', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.232513] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 906.235902] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 906.237317] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99a39f5c-57aa-47b6-ac59-f2904c3c2d82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.262643] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.262643] env[70013]: value = "task-4231007" [ 906.262643] env[70013]: _type = "Task" [ 906.262643] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.274150] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231007, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.293693] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231005, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.445293] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "39557c50-866a-4510-b840-b1a6a3e3890e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.445583] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "39557c50-866a-4510-b840-b1a6a3e3890e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.457118] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231003, 'name': ReconfigVM_Task, 'duration_secs': 0.572899} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.468308] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Reconfigured VM instance instance-00000037 to attach disk [datastore2] d84d8426-b3fd-4f0d-8410-0ee24fedb35d/d84d8426-b3fd-4f0d-8410-0ee24fedb35d.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.472032] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c73c01a-88c9-46a3-a1c3-443817fcdf34 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.481888] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 906.481888] env[70013]: value = "task-4231008" [ 906.481888] env[70013]: _type = "Task" [ 906.481888] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.498657] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231008, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.685059] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810998ff-cfae-4b25-80c6-c832f6f22744 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.706664] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedf4783-6689-49c7-8378-ff346f861d8f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.710166] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a8b0f4-2eba-a6c6-7197-a131e65b82a5, 'name': SearchDatastore_Task, 'duration_secs': 0.010788} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.711502] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6848f798-968e-4b5e-a8e0-6ba383baf81f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.715453] env[70013]: DEBUG nova.network.neutron [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Updated VIF entry in instance network info cache for port 7825762a-5525-4c92-8146-1ac33ba295f6. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.715637] env[70013]: DEBUG nova.network.neutron [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Updating instance_info_cache with network_info: [{"id": "7825762a-5525-4c92-8146-1ac33ba295f6", "address": "fa:16:3e:17:35:f5", "network": {"id": "f0029e8a-cbc6-4c49-b026-38d3aa38ee1c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-216013226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4c67d4c6cfb40d08ae8ee115ade7273", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7825762a-55", "ovs_interfaceid": "7825762a-5525-4c92-8146-1ac33ba295f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.745428] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ee797d-8edb-4db3-8c5e-7265ccf9f00a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.750672] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 906.750672] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527240e9-3901-ab88-f3cb-5821a74d7b72" [ 906.750672] env[70013]: _type = "Task" [ 906.750672] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.760578] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65efb5a7-3805-43e0-adeb-da73f1489300 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.770347] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527240e9-3901-ab88-f3cb-5821a74d7b72, 'name': SearchDatastore_Task, 'duration_secs': 0.011597} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.773991] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.774309] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. {{(pid=70013) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 906.774598] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdb29893-1f81-406b-9007-7200af9db23e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.785694] env[70013]: DEBUG nova.compute.provider_tree [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.793793] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231007, 'name': CreateVM_Task, 'duration_secs': 0.357235} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.795346] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.801024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.801024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.801024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 906.801024] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 906.801024] env[70013]: value = "task-4231009" [ 906.801024] env[70013]: _type = "Task" [ 906.801024] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.803651] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d459de3-40c5-41e4-9a85-812b6b0a6d99 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.806520] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231005, 'name': CreateSnapshot_Task, 'duration_secs': 0.740088} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.810885] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 906.812353] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101578a2-093d-42a9-a81a-d0b7a1826dd3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.817897] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 906.817897] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a06101-08c1-8b1a-8a42-0a141b859d53" [ 906.817897] env[70013]: _type = "Task" [ 906.817897] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.829672] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.840272] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a06101-08c1-8b1a-8a42-0a141b859d53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.999710] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231008, 'name': Rename_Task, 'duration_secs': 0.19503} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.000062] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.000406] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5dbae2ec-757d-4152-a4a2-ee3c0e49d304 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.011072] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 907.011072] env[70013]: value = "task-4231010" [ 907.011072] env[70013]: _type = "Task" [ 907.011072] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.025894] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231010, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.245142] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] Releasing lock "refresh_cache-2aefaa91-c439-486a-8b19-c6f45f52583f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.245488] env[70013]: DEBUG nova.compute.manager [req-d9b2ac53-2b57-4abd-89f8-c377a1728814 req-9c03a476-260a-4182-a9ff-9fcca35870d2 service nova] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Received event network-vif-deleted-b1d1b3a0-6686-4faf-981e-6772c3a821dd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 907.299110] env[70013]: DEBUG nova.scheduler.client.report [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.321523] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231009, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.341917] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 907.342384] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a06101-08c1-8b1a-8a42-0a141b859d53, 'name': SearchDatastore_Task, 'duration_secs': 0.022021} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.343220] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6111f91a-f1e6-4e01-8d3d-b11e01835b39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.347273] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.347711] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.348051] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.348247] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 907.348620] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.349125] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5758b72d-67e8-431d-9704-663973b52ec8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.361031] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 907.361031] env[70013]: value = "task-4231011" [ 907.361031] env[70013]: _type = "Task" [ 907.361031] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.362077] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.362181] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.363288] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-119b0ada-367e-4464-ae4c-389e3e279afc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.369797] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231011, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.374481] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 907.374481] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529284f6-a04d-0fb3-0225-f3aa951301f7" [ 907.374481] env[70013]: _type = "Task" [ 907.374481] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.383697] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529284f6-a04d-0fb3-0225-f3aa951301f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.526435] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231010, 'name': PowerOnVM_Task} progress is 87%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.808142] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.293s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.809739] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.963s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 907.813077] env[70013]: INFO nova.compute.claims [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.826922] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547906} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.827484] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. [ 907.828490] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a971c5a5-be9a-4989-b8b0-0f1f53b3fd22 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.857608] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.859927] env[70013]: INFO nova.scheduler.client.report [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Deleted allocations for instance be9a565b-c74c-442e-8087-cc4579b1c8e7 [ 907.860831] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11672654-587d-4411-bb81-7413d437a2c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.896158] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231011, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.900711] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529284f6-a04d-0fb3-0225-f3aa951301f7, 'name': SearchDatastore_Task, 'duration_secs': 0.010597} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.900981] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 907.900981] env[70013]: value = "task-4231012" [ 907.900981] env[70013]: _type = "Task" [ 907.900981] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.901921] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-323c3559-09bc-4ac8-8c24-b62c74585dc9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.911140] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 907.911140] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52eae043-b399-0c19-d9c6-cbe3ecfdc806" [ 907.911140] env[70013]: _type = "Task" [ 907.911140] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.916710] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231012, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.928080] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52eae043-b399-0c19-d9c6-cbe3ecfdc806, 'name': SearchDatastore_Task, 'duration_secs': 0.012526} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.928409] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.928687] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2aefaa91-c439-486a-8b19-c6f45f52583f/2aefaa91-c439-486a-8b19-c6f45f52583f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.928957] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78ad47cc-92e1-40b9-aa2a-8d4e480c366b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.937413] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 907.937413] env[70013]: value = "task-4231013" [ 907.937413] env[70013]: _type = "Task" [ 907.937413] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.948451] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.026905] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231010, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.382316] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231011, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.392070] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3be523cf-1ca4-49e2-a04c-a7e9a4ac871a tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "be9a565b-c74c-442e-8087-cc4579b1c8e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.936s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.423397] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231012, 'name': ReconfigVM_Task, 'duration_secs': 0.370784} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.424381] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.427191] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50e2ee3-d0c7-4a96-afec-2a87d397d86b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.472489] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1ba4203-7ef3-48bf-9bda-a75adda6210a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.491579] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520367} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.495024] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2aefaa91-c439-486a-8b19-c6f45f52583f/2aefaa91-c439-486a-8b19-c6f45f52583f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.495024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.495024] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 908.495024] env[70013]: value = "task-4231014" [ 908.495024] env[70013]: _type = "Task" [ 908.495024] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.495024] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-848908a6-24f9-4d07-9671-95df70b0f39b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.505385] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231014, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.507089] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 908.507089] env[70013]: value = "task-4231015" [ 908.507089] env[70013]: _type = "Task" [ 908.507089] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.527385] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231010, 'name': PowerOnVM_Task, 'duration_secs': 1.029751} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.527657] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.527917] env[70013]: INFO nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Took 7.95 seconds to spawn the instance on the hypervisor. [ 908.528157] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.528987] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8bd87b4-b733-4478-a558-9714fe25d7c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.876581] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231011, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.005814] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231014, 'name': ReconfigVM_Task, 'duration_secs': 0.341241} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.009716] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 909.010401] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a656c7fb-2d91-474e-926b-54a7df64902b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.025790] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.184514} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.026907] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.027298] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 909.027298] env[70013]: value = "task-4231017" [ 909.027298] env[70013]: _type = "Task" [ 909.027298] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.028040] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2401e92c-7438-44be-b723-c337d8efe1e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.051030] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231017, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.077092] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 2aefaa91-c439-486a-8b19-c6f45f52583f/2aefaa91-c439-486a-8b19-c6f45f52583f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.077397] env[70013]: INFO nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Took 46.60 seconds to build instance. [ 909.079804] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a811f99-0454-4c7b-bf25-7e493ee7ee44 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.100991] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 909.100991] env[70013]: value = "task-4231018" [ 909.100991] env[70013]: _type = "Task" [ 909.100991] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.113934] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231018, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.381507] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231011, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.400563] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334b2531-0511-4e82-935e-4d474d5bab75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.409905] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5729f2-f418-416a-abf3-e022374bbaee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.443888] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e370fe-86ee-4de3-ae78-32b154521947 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.456580] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61365e7f-613f-4ed5-9213-e14f2c46b656 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.475226] env[70013]: DEBUG nova.compute.provider_tree [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 909.541402] env[70013]: DEBUG oslo_vmware.api [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231017, 'name': PowerOnVM_Task, 'duration_secs': 0.482127} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.541792] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.545037] env[70013]: DEBUG nova.compute.manager [None req-7720d2ab-1ce9-468c-870a-62a67b16b4ea tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 909.546114] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b97fae1-b58b-4092-a243-9a1fc61e589c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.595550] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.314s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.614563] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231018, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.880378] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231011, 'name': CloneVM_Task, 'duration_secs': 2.375451} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.880691] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Created linked-clone VM from snapshot [ 909.881609] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ad2444-d4ed-4eea-be2e-606919a82372 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.895837] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Uploading image ca4066da-1b0a-465f-b38a-e1e6b704308d {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 909.917575] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 909.917575] env[70013]: value = "vm-837139" [ 909.917575] env[70013]: _type = "VirtualMachine" [ 909.917575] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 909.917869] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4de5424b-be3f-40e6-9e82-06e19537444e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.926350] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease: (returnval){ [ 909.926350] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e2c54-3127-844a-bccc-a04c0de6b311" [ 909.926350] env[70013]: _type = "HttpNfcLease" [ 909.926350] env[70013]: } obtained for exporting VM: (result){ [ 909.926350] env[70013]: value = "vm-837139" [ 909.926350] env[70013]: _type = "VirtualMachine" [ 909.926350] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 909.926651] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the lease: (returnval){ [ 909.926651] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e2c54-3127-844a-bccc-a04c0de6b311" [ 909.926651] env[70013]: _type = "HttpNfcLease" [ 909.926651] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 909.933812] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.933812] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e2c54-3127-844a-bccc-a04c0de6b311" [ 909.933812] env[70013]: _type = "HttpNfcLease" [ 909.933812] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 910.005916] env[70013]: ERROR nova.scheduler.client.report [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [req-f5348138-1f80-41e9-a33e-5982ad9586fe] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f5348138-1f80-41e9-a33e-5982ad9586fe"}]} [ 910.006270] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.006921] env[70013]: ERROR nova.compute.manager [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Traceback (most recent call last): [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] yield [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] self.set_inventory_for_provider( [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f5348138-1f80-41e9-a33e-5982ad9586fe"}]} [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] During handling of the above exception, another exception occurred: [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Traceback (most recent call last): [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] with self.rt.instance_claim(context, instance, node, allocs, [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] return f(*args, **kwargs) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] self._update(elevated, cn) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] self._update_to_placement(context, compute_node, startup) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] return attempt.get(self._wrap_exception) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] six.reraise(self.value[0], self.value[1], self.value[2]) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] raise value [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] self.reportclient.update_from_provider_tree( [ 910.006921] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 910.008860] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] with catch_all(pd.uuid): [ 910.008860] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 910.008860] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] self.gen.throw(typ, value, traceback) [ 910.008860] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 910.008860] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] raise exception.ResourceProviderSyncFailed() [ 910.008860] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 910.008860] env[70013]: ERROR nova.compute.manager [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] [ 910.008860] env[70013]: DEBUG nova.compute.utils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 910.009971] env[70013]: DEBUG nova.compute.manager [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Build of instance b853f0c4-b79e-4e30-8ba1-3d3fb7041809 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 910.011118] env[70013]: DEBUG nova.compute.manager [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 910.011118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "refresh_cache-b853f0c4-b79e-4e30-8ba1-3d3fb7041809" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.011118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired lock "refresh_cache-b853f0c4-b79e-4e30-8ba1-3d3fb7041809" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 910.011118] env[70013]: DEBUG nova.network.neutron [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 910.013025] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.011s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.014794] env[70013]: INFO nova.compute.claims [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.098776] env[70013]: DEBUG nova.compute.manager [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 910.115273] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231018, 'name': ReconfigVM_Task, 'duration_secs': 0.69854} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.115273] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 2aefaa91-c439-486a-8b19-c6f45f52583f/2aefaa91-c439-486a-8b19-c6f45f52583f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.115785] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff116e25-1e73-4a35-9d78-edd041fd6e60 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.123352] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 910.123352] env[70013]: value = "task-4231020" [ 910.123352] env[70013]: _type = "Task" [ 910.123352] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.133102] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231020, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.435314] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.435314] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e2c54-3127-844a-bccc-a04c0de6b311" [ 910.435314] env[70013]: _type = "HttpNfcLease" [ 910.435314] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 910.435693] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 910.435693] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e2c54-3127-844a-bccc-a04c0de6b311" [ 910.435693] env[70013]: _type = "HttpNfcLease" [ 910.435693] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 910.436511] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7166696e-94f2-4daf-96f8-c7e839477275 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.445790] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b29089-06ca-6c21-4e79-25076cf53a53/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 910.445982] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b29089-06ca-6c21-4e79-25076cf53a53/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 910.535391] env[70013]: DEBUG nova.network.neutron [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.554829] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fe62089d-e999-4096-8660-66be0cee500b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.635892] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231020, 'name': Rename_Task, 'duration_secs': 0.244948} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.636964] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.637296] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.637548] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c61c8cb1-359e-4f38-b8fd-8bcaa617bd05 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.648609] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 910.648609] env[70013]: value = "task-4231021" [ 910.648609] env[70013]: _type = "Task" [ 910.648609] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.658589] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231021, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.676934] env[70013]: DEBUG nova.network.neutron [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.048883] env[70013]: DEBUG nova.scheduler.client.report [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 911.066107] env[70013]: DEBUG nova.scheduler.client.report [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 911.066767] env[70013]: DEBUG nova.compute.provider_tree [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.082311] env[70013]: DEBUG nova.scheduler.client.report [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 911.108056] env[70013]: DEBUG nova.scheduler.client.report [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 911.165376] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231021, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.182112] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Releasing lock "refresh_cache-b853f0c4-b79e-4e30-8ba1-3d3fb7041809" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.182413] env[70013]: DEBUG nova.compute.manager [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 911.182712] env[70013]: DEBUG nova.compute.manager [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 911.182965] env[70013]: DEBUG nova.network.neutron [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.218663] env[70013]: DEBUG nova.network.neutron [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 911.665186] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231021, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.672149] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7bfbe20-ce03-4dc0-a58c-74ee827ec7ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.681256] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e5bb2c-e469-4c45-ab07-2ed17b3df187 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.730696] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.731103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.731393] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.731678] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.731919] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.734785] env[70013]: INFO nova.compute.manager [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Terminating instance [ 911.736884] env[70013]: DEBUG nova.network.neutron [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.744231] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f57f525-71ca-4374-ad28-8852f0ccba8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.754120] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74d3340-5a40-41f6-a648-3f1613b08ec1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.777347] env[70013]: DEBUG nova.compute.provider_tree [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.966257] env[70013]: INFO nova.compute.manager [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Unrescuing [ 911.966723] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.967034] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.967342] env[70013]: DEBUG nova.network.neutron [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.169635] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231021, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.240436] env[70013]: INFO nova.compute.manager [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: b853f0c4-b79e-4e30-8ba1-3d3fb7041809] Took 1.06 seconds to deallocate network for instance. [ 912.246748] env[70013]: DEBUG nova.compute.manager [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 912.246748] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.247632] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a11ec4-6938-4655-8de4-19549b145063 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.256866] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.257779] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe2ff1c7-fb9d-4667-9be3-06af1ea2c95d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.267276] env[70013]: DEBUG oslo_vmware.api [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 912.267276] env[70013]: value = "task-4231022" [ 912.267276] env[70013]: _type = "Task" [ 912.267276] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.283105] env[70013]: DEBUG oslo_vmware.api [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4231022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.305134] env[70013]: ERROR nova.scheduler.client.report [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [req-9293afa6-3bc3-4fda-ab1f-545cfaef4bb0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9293afa6-3bc3-4fda-ab1f-545cfaef4bb0"}]} [ 912.305551] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.306221] env[70013]: ERROR nova.compute.manager [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Traceback (most recent call last): [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] yield [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] self.set_inventory_for_provider( [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9293afa6-3bc3-4fda-ab1f-545cfaef4bb0"}]} [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] During handling of the above exception, another exception occurred: [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Traceback (most recent call last): [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] with self.rt.instance_claim(context, instance, node, allocs, [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] return f(*args, **kwargs) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] self._update(elevated, cn) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] self._update_to_placement(context, compute_node, startup) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] return attempt.get(self._wrap_exception) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] six.reraise(self.value[0], self.value[1], self.value[2]) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] raise value [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] self.reportclient.update_from_provider_tree( [ 912.306221] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 912.308741] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] with catch_all(pd.uuid): [ 912.308741] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 912.308741] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] self.gen.throw(typ, value, traceback) [ 912.308741] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 912.308741] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] raise exception.ResourceProviderSyncFailed() [ 912.308741] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 912.308741] env[70013]: ERROR nova.compute.manager [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] [ 912.308741] env[70013]: DEBUG nova.compute.utils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 912.308741] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.103s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.308741] env[70013]: DEBUG nova.objects.instance [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lazy-loading 'resources' on Instance uuid 3b29c99a-0d55-40b1-a155-199b1ced2146 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.310581] env[70013]: DEBUG nova.compute.manager [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Build of instance cfccb245-dec8-4dec-bd73-33fe18ec8a19 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 912.310581] env[70013]: DEBUG nova.compute.manager [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 912.310765] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Acquiring lock "refresh_cache-cfccb245-dec8-4dec-bd73-33fe18ec8a19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.310914] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Acquired lock "refresh_cache-cfccb245-dec8-4dec-bd73-33fe18ec8a19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.311019] env[70013]: DEBUG nova.network.neutron [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.668332] env[70013]: DEBUG oslo_vmware.api [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231021, 'name': PowerOnVM_Task, 'duration_secs': 1.932835} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.668710] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.668855] env[70013]: INFO nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Took 9.47 seconds to spawn the instance on the hypervisor. [ 912.669048] env[70013]: DEBUG nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 912.669863] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25838697-2093-4151-a5ee-fa018db73eed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.780528] env[70013]: DEBUG oslo_vmware.api [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4231022, 'name': PowerOffVM_Task, 'duration_secs': 0.301693} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.780849] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.781441] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 912.781734] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-373f863d-f2c3-4b32-824d-75fdd2d9c066 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.799666] env[70013]: DEBUG nova.network.neutron [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Updating instance_info_cache with network_info: [{"id": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "address": "fa:16:3e:01:6f:a2", "network": {"id": "cf3c83ee-c081-4f73-8ac2-5ddaff2ef56a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1948618544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "208376e0d950450cb2ce841e2d049cd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a9242b0-4a", "ovs_interfaceid": "4a9242b0-4afa-443e-9ccd-dc35313fc59a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.836807] env[70013]: DEBUG nova.network.neutron [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.848604] env[70013]: DEBUG nova.scheduler.client.report [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 912.863509] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.863769] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.863927] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Deleting the datastore file [datastore2] 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.864243] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-797b6196-7931-48ba-b6da-c7310529fc4b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.873159] env[70013]: DEBUG oslo_vmware.api [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for the task: (returnval){ [ 912.873159] env[70013]: value = "task-4231024" [ 912.873159] env[70013]: _type = "Task" [ 912.873159] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.874324] env[70013]: DEBUG nova.scheduler.client.report [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 912.874539] env[70013]: DEBUG nova.compute.provider_tree [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.888471] env[70013]: DEBUG oslo_vmware.api [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4231024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.894018] env[70013]: DEBUG nova.scheduler.client.report [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 912.929469] env[70013]: DEBUG nova.scheduler.client.report [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 912.964143] env[70013]: DEBUG nova.network.neutron [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.195261] env[70013]: INFO nova.compute.manager [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Took 47.77 seconds to build instance. [ 913.291265] env[70013]: INFO nova.scheduler.client.report [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Deleted allocations for instance b853f0c4-b79e-4e30-8ba1-3d3fb7041809 [ 913.307091] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "refresh_cache-789fffd9-6725-4bf6-9144-dd603b0a521f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 913.308851] env[70013]: DEBUG nova.objects.instance [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lazy-loading 'flavor' on Instance uuid 789fffd9-6725-4bf6-9144-dd603b0a521f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.393127] env[70013]: DEBUG oslo_vmware.api [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Task: {'id': task-4231024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389111} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.394473] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.394687] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.394872] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.395071] env[70013]: INFO nova.compute.manager [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 913.395347] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 913.399418] env[70013]: DEBUG nova.compute.manager [-] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 913.399630] env[70013]: DEBUG nova.network.neutron [-] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.468141] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Releasing lock "refresh_cache-cfccb245-dec8-4dec-bd73-33fe18ec8a19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 913.468966] env[70013]: DEBUG nova.compute.manager [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 913.469247] env[70013]: DEBUG nova.compute.manager [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 913.471240] env[70013]: DEBUG nova.network.neutron [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.498708] env[70013]: DEBUG nova.network.neutron [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 913.520391] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b283b8a8-cf41-46d1-860a-47de7f0090ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.537567] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076b3bee-9c0e-40ee-be26-ac1bc3ac8435 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.584848] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f664e97-e320-4d7f-b717-5ff306495257 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.593565] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719f0ede-5bca-43bb-91a4-21ea3ca42277 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.612152] env[70013]: DEBUG nova.compute.provider_tree [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.699076] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6d076486-1029-4d06-98fb-9a46b1764b81 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.378s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.736047] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.736047] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.736331] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.736530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.736705] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.740147] env[70013]: INFO nova.compute.manager [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Terminating instance [ 913.800645] env[70013]: DEBUG oslo_concurrency.lockutils [None req-27e0e80b-683c-4b0b-b905-58fd6b312000 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "b853f0c4-b79e-4e30-8ba1-3d3fb7041809" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.738s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.817889] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee0cb1a-e658-4e5c-baac-67b6ed8f0761 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.842311] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.842738] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7ce1151-6f89-4f32-9b8b-9e651cc524cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.852200] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 913.852200] env[70013]: value = "task-4231025" [ 913.852200] env[70013]: _type = "Task" [ 913.852200] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.861723] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231025, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.900064] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "2aefaa91-c439-486a-8b19-c6f45f52583f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.901035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.901035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "2aefaa91-c439-486a-8b19-c6f45f52583f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.901035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.901035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.905174] env[70013]: INFO nova.compute.manager [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Terminating instance [ 914.003014] env[70013]: DEBUG nova.network.neutron [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.116710] env[70013]: DEBUG nova.scheduler.client.report [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 914.172688] env[70013]: DEBUG nova.compute.manager [req-cf4a4a94-70e6-490a-934e-55019f207feb req-27d51da1-02b3-4248-b1dd-ed732c8e0181 service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Received event network-vif-deleted-0da6d475-040d-43e8-9139-88652285a5fe {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 914.173227] env[70013]: INFO nova.compute.manager [req-cf4a4a94-70e6-490a-934e-55019f207feb req-27d51da1-02b3-4248-b1dd-ed732c8e0181 service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Neutron deleted interface 0da6d475-040d-43e8-9139-88652285a5fe; detaching it from the instance and deleting it from the info cache [ 914.173924] env[70013]: DEBUG nova.network.neutron [req-cf4a4a94-70e6-490a-934e-55019f207feb req-27d51da1-02b3-4248-b1dd-ed732c8e0181 service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.202629] env[70013]: DEBUG nova.compute.manager [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 914.245736] env[70013]: DEBUG nova.compute.manager [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 914.245963] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.247479] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395a84a3-8824-4fcd-8264-2b6b57e3a7d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.262664] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.262664] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-360df6d7-ec30-4191-bd3f-d9b5f07df65d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.270816] env[70013]: DEBUG oslo_vmware.api [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 914.270816] env[70013]: value = "task-4231026" [ 914.270816] env[70013]: _type = "Task" [ 914.270816] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.284701] env[70013]: DEBUG oslo_vmware.api [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231026, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.303266] env[70013]: DEBUG nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 914.371909] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231025, 'name': PowerOffVM_Task, 'duration_secs': 0.370101} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.372231] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.377768] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Reconfiguring VM instance instance-00000032 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 914.378886] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bed72bbb-7117-4fc6-baf0-e1a6948a0007 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.407331] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 914.407331] env[70013]: value = "task-4231027" [ 914.407331] env[70013]: _type = "Task" [ 914.407331] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.412850] env[70013]: DEBUG nova.compute.manager [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 914.413114] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.414029] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2007a9-8849-4b34-bac3-6bcf4ec34d70 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.424834] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231027, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.427600] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.428139] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c2f2064-7c00-459f-8f84-3dcc7541b3c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.436569] env[70013]: DEBUG oslo_vmware.api [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 914.436569] env[70013]: value = "task-4231028" [ 914.436569] env[70013]: _type = "Task" [ 914.436569] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.450193] env[70013]: DEBUG oslo_vmware.api [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231028, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.508307] env[70013]: INFO nova.compute.manager [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] [instance: cfccb245-dec8-4dec-bd73-33fe18ec8a19] Took 1.04 seconds to deallocate network for instance. [ 914.594266] env[70013]: DEBUG nova.network.neutron [-] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.624991] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.316s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.627823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.942s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.628521] env[70013]: DEBUG nova.objects.instance [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lazy-loading 'resources' on Instance uuid 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.656744] env[70013]: INFO nova.scheduler.client.report [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Deleted allocations for instance 3b29c99a-0d55-40b1-a155-199b1ced2146 [ 914.682043] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ecd5c99-7760-4a48-bd2b-8f21882f5ae3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.691638] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5fbab1-3a14-4667-9dce-abedee1f608f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.731595] env[70013]: DEBUG nova.compute.manager [req-cf4a4a94-70e6-490a-934e-55019f207feb req-27d51da1-02b3-4248-b1dd-ed732c8e0181 service nova] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Detach interface failed, port_id=0da6d475-040d-43e8-9139-88652285a5fe, reason: Instance 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 914.735155] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.783060] env[70013]: DEBUG oslo_vmware.api [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231026, 'name': PowerOffVM_Task, 'duration_secs': 0.506305} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.783434] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.783660] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.783960] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d8d91f6-16f2-400d-86ee-9413989b1fd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.829517] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.871724] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.871944] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.873486] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Deleting the datastore file [datastore2] d84d8426-b3fd-4f0d-8410-0ee24fedb35d {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.873486] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52baa953-44f6-44a3-ba07-eecd3ca7be69 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.880638] env[70013]: DEBUG oslo_vmware.api [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 914.880638] env[70013]: value = "task-4231030" [ 914.880638] env[70013]: _type = "Task" [ 914.880638] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.894368] env[70013]: DEBUG oslo_vmware.api [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.923995] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231027, 'name': ReconfigVM_Task, 'duration_secs': 0.343616} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.924666] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Reconfigured VM instance instance-00000032 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 914.924666] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.928653] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37cee556-0a27-4b89-a6c8-d096002cb3da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.936091] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 914.936091] env[70013]: value = "task-4231031" [ 914.936091] env[70013]: _type = "Task" [ 914.936091] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.948814] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231031, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.952974] env[70013]: DEBUG oslo_vmware.api [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231028, 'name': PowerOffVM_Task, 'duration_secs': 0.347837} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.953151] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.954037] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.954037] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6704632b-2471-4d73-8484-ef395c9bde7b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.040884] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 915.041297] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 915.041530] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Deleting the datastore file [datastore2] 2aefaa91-c439-486a-8b19-c6f45f52583f {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.043975] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c302316f-bdab-4cc3-a223-d04dc57f7a6c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.054879] env[70013]: DEBUG oslo_vmware.api [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for the task: (returnval){ [ 915.054879] env[70013]: value = "task-4231033" [ 915.054879] env[70013]: _type = "Task" [ 915.054879] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.067049] env[70013]: DEBUG oslo_vmware.api [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231033, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.101211] env[70013]: INFO nova.compute.manager [-] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Took 1.70 seconds to deallocate network for instance. [ 915.164410] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7d3bd08b-ae51-49dc-b344-035d69a10fd0 tempest-MigrationsAdminTest-838255870 tempest-MigrationsAdminTest-838255870-project-member] Lock "3b29c99a-0d55-40b1-a155-199b1ced2146" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.544s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.395553] env[70013]: DEBUG oslo_vmware.api [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245626} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.395968] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.396280] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.396582] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.396864] env[70013]: INFO nova.compute.manager [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 915.397278] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 915.401334] env[70013]: DEBUG nova.compute.manager [-] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 915.401445] env[70013]: DEBUG nova.network.neutron [-] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 915.452893] env[70013]: DEBUG oslo_vmware.api [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231031, 'name': PowerOnVM_Task, 'duration_secs': 0.430263} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.452893] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.453179] env[70013]: DEBUG nova.compute.manager [None req-6e5cb4a5-88c2-40dc-99cd-983950958cef tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 915.453807] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c189d2ef-425d-4cab-8578-b0c3235274c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.552810] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.553078] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.556355] env[70013]: INFO nova.scheduler.client.report [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Deleted allocations for instance cfccb245-dec8-4dec-bd73-33fe18ec8a19 [ 915.573586] env[70013]: DEBUG oslo_vmware.api [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Task: {'id': task-4231033, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156768} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.573916] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.574123] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.574318] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.574480] env[70013]: INFO nova.compute.manager [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Took 1.16 seconds to destroy the instance on the hypervisor. [ 915.574727] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 915.577570] env[70013]: DEBUG nova.compute.manager [-] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 915.577705] env[70013]: DEBUG nova.network.neutron [-] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 915.611538] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.640734] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0574e73-b9e1-4630-bf95-ecbcc14011b0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.655219] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b18745-7e89-4b09-bcac-926955974d62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.690777] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f42d3de-bdd8-4e4d-95b9-e7e336ed2d24 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.700661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c712f27-8c27-42a0-9cf5-d3c8e0ec1565 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.718652] env[70013]: DEBUG nova.compute.provider_tree [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.063139] env[70013]: INFO nova.compute.manager [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Detaching volume 8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0 [ 916.066974] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9294321-36ae-4c24-9550-361fd8b65709 tempest-ServersTestManualDisk-1390980805 tempest-ServersTestManualDisk-1390980805-project-member] Lock "cfccb245-dec8-4dec-bd73-33fe18ec8a19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.816s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.111318] env[70013]: INFO nova.virt.block_device [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Attempting to driver detach volume 8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0 from mountpoint /dev/sdb [ 916.111618] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 916.111890] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837110', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'name': 'volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2c0e4326-c33a-42bb-b793-a100157b1c03', 'attached_at': '', 'detached_at': '', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'serial': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 916.114228] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3110c56b-333c-4043-ba31-c22ddcfb05e9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.142378] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642d7fc2-148c-434d-97b4-eca29a0e4b03 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.150268] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdb48e0-adc8-464c-94e0-b86ec57e16d0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.178382] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72230ea-7cae-4833-bda6-fd45f153dcf4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.196725] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] The volume has not been displaced from its original location: [localhost-esx-install-datastore (2)] volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0/volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 916.202703] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Reconfiguring VM instance instance-00000017 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 916.203125] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcd6bb56-3c7a-418f-be79-93406488c128 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.224508] env[70013]: DEBUG nova.scheduler.client.report [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 916.230085] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 916.230085] env[70013]: value = "task-4231034" [ 916.230085] env[70013]: _type = "Task" [ 916.230085] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.241014] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231034, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.253354] env[70013]: DEBUG nova.compute.manager [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Received event network-vif-deleted-c7d197fd-939c-426a-b373-674e428f6655 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 916.253354] env[70013]: INFO nova.compute.manager [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Neutron deleted interface c7d197fd-939c-426a-b373-674e428f6655; detaching it from the instance and deleting it from the info cache [ 916.253701] env[70013]: DEBUG nova.network.neutron [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.326418] env[70013]: DEBUG nova.network.neutron [-] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.387528] env[70013]: DEBUG nova.network.neutron [-] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.737327] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.110s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.742548] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.827s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.742832] env[70013]: DEBUG nova.objects.instance [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lazy-loading 'resources' on Instance uuid 0b0b2769-74ba-4ba2-9e8b-148cc2976321 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.752123] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231034, 'name': ReconfigVM_Task, 'duration_secs': 0.310401} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.752914] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Reconfigured VM instance instance-00000017 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 916.758876] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2f91343-a04e-463f-87f3-9a44b262d1d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.772137] env[70013]: INFO nova.scheduler.client.report [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Deleted allocations for instance 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268 [ 916.776017] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3bedfee7-5a8c-4f7b-b798-d32cac816c48 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.789899] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 916.789899] env[70013]: value = "task-4231035" [ 916.789899] env[70013]: _type = "Task" [ 916.789899] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.794206] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb033b7d-9a78-45c9-9a37-a9019df9063d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.813689] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231035, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.830406] env[70013]: INFO nova.compute.manager [-] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Took 1.43 seconds to deallocate network for instance. [ 916.830917] env[70013]: DEBUG nova.compute.manager [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Detach interface failed, port_id=c7d197fd-939c-426a-b373-674e428f6655, reason: Instance d84d8426-b3fd-4f0d-8410-0ee24fedb35d could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 916.831482] env[70013]: DEBUG nova.compute.manager [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Received event network-vif-deleted-7825762a-5525-4c92-8146-1ac33ba295f6 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 916.832105] env[70013]: INFO nova.compute.manager [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Neutron deleted interface 7825762a-5525-4c92-8146-1ac33ba295f6; detaching it from the instance and deleting it from the info cache [ 916.832378] env[70013]: DEBUG nova.network.neutron [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.890940] env[70013]: INFO nova.compute.manager [-] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Took 1.31 seconds to deallocate network for instance. [ 917.283108] env[70013]: DEBUG oslo_concurrency.lockutils [None req-36443c68-0a2e-41b4-9ab5-ada1ebc4591b tempest-InstanceActionsTestJSON-1929193662 tempest-InstanceActionsTestJSON-1929193662-project-member] Lock "6b458a7e-f786-41f6-b8ac-9f6f4ddcb268" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.229s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.301373] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231035, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.337682] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d74242c5-fa54-46a1-bd5c-74f297815dc5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.341271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.348522] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff666e62-2e40-40de-a5e5-46efe999f3ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.384276] env[70013]: DEBUG nova.compute.manager [req-3588e647-1567-4fdb-a0f3-7cfbeba9afdb req-e592a5ff-6c98-4a2f-bc94-2474df7c4096 service nova] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Detach interface failed, port_id=7825762a-5525-4c92-8146-1ac33ba295f6, reason: Instance 2aefaa91-c439-486a-8b19-c6f45f52583f could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 917.398323] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.693303] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96caafc5-803d-4c1c-9f37-098dd060e748 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.707189] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f2adca-5453-4c86-a063-7ad5a7aa38ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.746796] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce480217-c833-4284-ab16-bc05a6f34d6e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.756702] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d17541-9351-4e19-88e4-3c0bcfd02dec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.772027] env[70013]: DEBUG nova.compute.provider_tree [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.802757] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231035, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.279020] env[70013]: DEBUG nova.scheduler.client.report [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 918.305034] env[70013]: DEBUG oslo_vmware.api [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231035, 'name': ReconfigVM_Task, 'duration_secs': 1.186591} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.305805] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837110', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'name': 'volume-8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2c0e4326-c33a-42bb-b793-a100157b1c03', 'attached_at': '', 'detached_at': '', 'volume_id': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0', 'serial': '8e80e49c-8b1f-47f0-a5fd-02e980b0cfe0'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 918.790993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.046s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 918.794350] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.684s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.797078] env[70013]: INFO nova.compute.claims [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.823198] env[70013]: INFO nova.scheduler.client.report [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Deleted allocations for instance 0b0b2769-74ba-4ba2-9e8b-148cc2976321 [ 918.881283] env[70013]: DEBUG nova.objects.instance [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lazy-loading 'flavor' on Instance uuid 2c0e4326-c33a-42bb-b793-a100157b1c03 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.136154] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b29089-06ca-6c21-4e79-25076cf53a53/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 919.138608] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f0383e-44df-4d24-bbc5-9b2e5d8be8c8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.148906] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b29089-06ca-6c21-4e79-25076cf53a53/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 919.149029] env[70013]: ERROR oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b29089-06ca-6c21-4e79-25076cf53a53/disk-0.vmdk due to incomplete transfer. [ 919.149293] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3d993449-2bb7-4687-b89f-d29d116046ed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.160922] env[70013]: DEBUG oslo_vmware.rw_handles [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b29089-06ca-6c21-4e79-25076cf53a53/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 919.161741] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Uploaded image ca4066da-1b0a-465f-b38a-e1e6b704308d to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 919.165722] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 919.165722] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-af07d9a3-9306-416a-bc24-e2470a0f1add {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.172114] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 919.172114] env[70013]: value = "task-4231036" [ 919.172114] env[70013]: _type = "Task" [ 919.172114] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.181506] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231036, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.332932] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70ba3fb0-9a8a-499e-9d3c-15ac27b9cc96 tempest-ServersAdminTestJSON-372282544 tempest-ServersAdminTestJSON-372282544-project-member] Lock "0b0b2769-74ba-4ba2-9e8b-148cc2976321" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.721s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.686495] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231036, 'name': Destroy_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.897679] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4eaf2ecf-8d5c-47e3-8ab3-cd5d4bd2de5f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.344s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.187964] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231036, 'name': Destroy_Task, 'duration_secs': 0.529274} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.188485] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Destroyed the VM [ 920.188878] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 920.191066] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f83cd3d8-9ef1-4bfa-8775-57351ac5d2e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.201600] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "8d28f827-c773-4f49-b89d-7d3dd8df4e70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.201963] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "8d28f827-c773-4f49-b89d-7d3dd8df4e70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.211917] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 920.211917] env[70013]: value = "task-4231037" [ 920.211917] env[70013]: _type = "Task" [ 920.211917] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.223096] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231037, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.324038] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f221b08b-8625-4f9a-8780-62340d2c0be0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.332837] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7253e140-fb80-43d1-9763-a9808ff60c9d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.371546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.371546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.371742] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.372033] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.372331] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.375267] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41cc9d3-4e33-475b-a153-0ea51aec1fb3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.378428] env[70013]: INFO nova.compute.manager [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Terminating instance [ 920.386929] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9cbd1e-1874-47da-b695-e69642921919 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.407804] env[70013]: DEBUG nova.compute.provider_tree [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.706338] env[70013]: DEBUG nova.compute.manager [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 920.726477] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231037, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.885402] env[70013]: DEBUG nova.compute.manager [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 920.885402] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.886352] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fce3054-f95c-4924-8243-e5f1e79f4a03 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.896430] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.897233] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b996696-032e-436c-9aec-5ea34792186d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.905198] env[70013]: DEBUG oslo_vmware.api [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 920.905198] env[70013]: value = "task-4231038" [ 920.905198] env[70013]: _type = "Task" [ 920.905198] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.911581] env[70013]: DEBUG nova.scheduler.client.report [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 920.922602] env[70013]: DEBUG oslo_vmware.api [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.233928] env[70013]: DEBUG oslo_vmware.api [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231037, 'name': RemoveSnapshot_Task, 'duration_secs': 0.683208} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.235654] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 921.235947] env[70013]: INFO nova.compute.manager [None req-30a85635-9011-43f8-b565-9ecf0c80cfcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Took 16.00 seconds to snapshot the instance on the hypervisor. [ 921.240947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.419344] env[70013]: DEBUG oslo_vmware.api [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231038, 'name': PowerOffVM_Task, 'duration_secs': 0.490334} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.421968] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.421968] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.421968] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2c14725-76e2-426f-852c-bc21fdb364a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.426867] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.633s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.427445] env[70013]: DEBUG nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 921.430438] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.119s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.436452] env[70013]: INFO nova.compute.claims [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.495780] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.497048] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.497581] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Deleting the datastore file [datastore2] 2c0e4326-c33a-42bb-b793-a100157b1c03 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.498366] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d74ca34-de5a-4b8b-b7d6-66fa4d7167d8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.510776] env[70013]: DEBUG oslo_vmware.api [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for the task: (returnval){ [ 921.510776] env[70013]: value = "task-4231040" [ 921.510776] env[70013]: _type = "Task" [ 921.510776] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.536067] env[70013]: DEBUG oslo_vmware.api [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.580445] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "8dbef727-13e9-4231-95a8-65d015ee13be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.580606] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.948256] env[70013]: DEBUG nova.compute.utils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 921.950784] env[70013]: DEBUG nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 922.023212] env[70013]: DEBUG oslo_vmware.api [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Task: {'id': task-4231040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169875} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.023654] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.023654] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.023823] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.024016] env[70013]: INFO nova.compute.manager [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Took 1.14 seconds to destroy the instance on the hypervisor. [ 922.024284] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 922.024486] env[70013]: DEBUG nova.compute.manager [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 922.024577] env[70013]: DEBUG nova.network.neutron [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.455253] env[70013]: DEBUG nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 923.047393] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94c95e5-114a-431c-9ecf-52080221edb3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.056785] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664782c5-21e1-466a-b603-347472785ddc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.093541] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c332c5dc-660c-4689-9d2e-5ca6139a10ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.104809] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e30955-0484-49a3-a1d4-fb80fffcb98a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.121329] env[70013]: DEBUG nova.compute.provider_tree [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.319683] env[70013]: DEBUG nova.compute.manager [req-93b8d31c-6daa-499c-ab80-9e1fc6c246ac req-1b30a393-bb8f-4315-bba7-edd49f016b36 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Received event network-vif-deleted-aab77ffc-40a3-47ea-abe4-1abb8298eba0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 923.319909] env[70013]: INFO nova.compute.manager [req-93b8d31c-6daa-499c-ab80-9e1fc6c246ac req-1b30a393-bb8f-4315-bba7-edd49f016b36 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Neutron deleted interface aab77ffc-40a3-47ea-abe4-1abb8298eba0; detaching it from the instance and deleting it from the info cache [ 923.320099] env[70013]: DEBUG nova.network.neutron [req-93b8d31c-6daa-499c-ab80-9e1fc6c246ac req-1b30a393-bb8f-4315-bba7-edd49f016b36 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.471344] env[70013]: DEBUG nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 923.514006] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 923.514940] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.515333] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 923.515751] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.515992] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 923.516248] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 923.516671] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 923.516925] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 923.517333] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 923.517843] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 923.518429] env[70013]: DEBUG nova.virt.hardware [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 923.519478] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31235b2-7d92-4776-9206-10f414fc7185 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.532567] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1409bf31-6b3a-441a-9159-344ec16621a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.551482] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.560344] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Creating folder: Project (7ddb0bea6af04354b0ce6a878520639b). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.560344] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66353871-72eb-4e79-995a-afb97a93e80a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.574724] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Created folder: Project (7ddb0bea6af04354b0ce6a878520639b) in parent group-v836999. [ 923.574724] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Creating folder: Instances. Parent ref: group-v837140. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.574724] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95e0be06-fe7c-48c8-9350-6d86667692ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.586178] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Created folder: Instances in parent group-v837140. [ 923.587363] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 923.587694] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.588506] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55d8131a-49da-457f-8a9e-42779f9b763b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.608073] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.608073] env[70013]: value = "task-4231043" [ 923.608073] env[70013]: _type = "Task" [ 923.608073] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.617384] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231043, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.625963] env[70013]: DEBUG nova.scheduler.client.report [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 923.769256] env[70013]: DEBUG nova.network.neutron [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.822593] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10da117c-74e3-4d65-b09f-2e202f2ace9b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.834727] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43637d56-c3da-4546-9df9-4527b226f1f2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.868336] env[70013]: DEBUG nova.compute.manager [req-93b8d31c-6daa-499c-ab80-9e1fc6c246ac req-1b30a393-bb8f-4315-bba7-edd49f016b36 service nova] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Detach interface failed, port_id=aab77ffc-40a3-47ea-abe4-1abb8298eba0, reason: Instance 2c0e4326-c33a-42bb-b793-a100157b1c03 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 924.121513] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231043, 'name': CreateVM_Task, 'duration_secs': 0.299483} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.122292] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.122785] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.122948] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 924.123303] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 924.123572] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bb09298-3a6d-4622-9a23-dce169ac5463 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.132164] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 924.132164] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526d74df-0816-7e22-fbe8-75ba2ff696ea" [ 924.132164] env[70013]: _type = "Task" [ 924.132164] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.139447] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.706s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 924.139447] env[70013]: DEBUG nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 924.143943] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.810s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.146108] env[70013]: INFO nova.compute.claims [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 924.156898] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526d74df-0816-7e22-fbe8-75ba2ff696ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.276191] env[70013]: INFO nova.compute.manager [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Took 2.25 seconds to deallocate network for instance. [ 924.644439] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526d74df-0816-7e22-fbe8-75ba2ff696ea, 'name': SearchDatastore_Task, 'duration_secs': 0.040363} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.644776] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 924.645066] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.648490] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.648490] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 924.648490] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.648490] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a04b358-bc78-45d7-8faa-a85a25b9d617 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.653335] env[70013]: DEBUG nova.compute.utils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 924.655153] env[70013]: DEBUG nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 924.655153] env[70013]: DEBUG nova.network.neutron [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.660262] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.660262] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.660262] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-750b3d6c-933f-429b-92d7-240a15817bae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.666997] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 924.666997] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5244b10f-8d8a-b90d-b11c-c404c8f594bd" [ 924.666997] env[70013]: _type = "Task" [ 924.666997] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.688022] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5244b10f-8d8a-b90d-b11c-c404c8f594bd, 'name': SearchDatastore_Task, 'duration_secs': 0.011545} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.688645] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c630369-e836-4924-b6db-54ad7db25ceb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.695690] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 924.695690] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5297e661-a774-2907-150e-bbdc7c3fbf40" [ 924.695690] env[70013]: _type = "Task" [ 924.695690] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.706322] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5297e661-a774-2907-150e-bbdc7c3fbf40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.729224] env[70013]: DEBUG nova.policy [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3bcb0d0bd1e44eb960d676070189f27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3d42e73dc0a4a06a2022d8b54e13e12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 924.784474] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.140017] env[70013]: DEBUG nova.network.neutron [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Successfully created port: 72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.167174] env[70013]: DEBUG nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 925.217977] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5297e661-a774-2907-150e-bbdc7c3fbf40, 'name': SearchDatastore_Task, 'duration_secs': 0.030025} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.217977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.218868] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 925.223381] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af3f603d-73ff-46b5-a14b-521ffd0b0c82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.233412] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 925.233412] env[70013]: value = "task-4231044" [ 925.233412] env[70013]: _type = "Task" [ 925.233412] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.246320] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.725725] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442423ef-d84b-4e85-b451-0b71f0d15e1d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.738854] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905b629b-ef42-4b32-901a-ef6e1abc3a32 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.749448] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231044, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.779025] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256278b8-db82-48c9-b045-c4aa07ba9101 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.787532] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a9ad59-3c05-4f82-962e-ea1eec24a5c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.804101] env[70013]: DEBUG nova.compute.provider_tree [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.179111] env[70013]: DEBUG nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 926.207366] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 926.207608] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.207759] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 926.207941] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.208103] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 926.208249] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 926.208451] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 926.208606] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 926.208764] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 926.208921] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 926.209103] env[70013]: DEBUG nova.virt.hardware [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 926.210014] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2d775c-c87f-4631-8c2e-76681aea45f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.224754] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd8dc02-9950-4861-869b-6448b09285d0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.252425] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517581} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.252703] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.252914] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.253232] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8bec10fd-59b3-48fd-bfaf-32a5c4bd61b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.260961] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 926.260961] env[70013]: value = "task-4231045" [ 926.260961] env[70013]: _type = "Task" [ 926.260961] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.273733] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231045, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.307600] env[70013]: DEBUG nova.scheduler.client.report [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 926.774289] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231045, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074912} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.774289] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.774380] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a0e806-45ca-4736-92fa-26aba4b35d7c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.799945] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.800361] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49566d8d-1ca4-4591-8c07-d24ae8b0eec8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.819702] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.820286] env[70013]: DEBUG nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 926.825766] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.408s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.825766] env[70013]: DEBUG nova.objects.instance [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lazy-loading 'resources' on Instance uuid a53389ce-d0c2-48ae-83e3-ee58fbdbfb01 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.831059] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 926.831059] env[70013]: value = "task-4231046" [ 926.831059] env[70013]: _type = "Task" [ 926.831059] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.840555] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231046, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.285599] env[70013]: DEBUG nova.network.neutron [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Successfully updated port: 72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.331326] env[70013]: DEBUG nova.compute.utils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 927.340281] env[70013]: DEBUG nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 927.340281] env[70013]: DEBUG nova.network.neutron [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 927.351525] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231046, 'name': ReconfigVM_Task, 'duration_secs': 0.349333} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.352766] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.353612] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3aa5c8a9-ff03-404f-9b0c-c424c0b77919 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.363774] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 927.363774] env[70013]: value = "task-4231047" [ 927.363774] env[70013]: _type = "Task" [ 927.363774] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.376759] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231047, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.388726] env[70013]: DEBUG nova.policy [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1634fc5b4baa40c6bd9e3bef9eb77864', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f5b0bd88f5046f098579c5d59e3f0c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 927.457621] env[70013]: DEBUG nova.compute.manager [req-5023e738-303e-4607-92c6-453a219a98e8 req-bcad1944-f21a-4c19-92d0-96ef2453d6e5 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Received event network-vif-plugged-72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 927.457854] env[70013]: DEBUG oslo_concurrency.lockutils [req-5023e738-303e-4607-92c6-453a219a98e8 req-bcad1944-f21a-4c19-92d0-96ef2453d6e5 service nova] Acquiring lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.458082] env[70013]: DEBUG oslo_concurrency.lockutils [req-5023e738-303e-4607-92c6-453a219a98e8 req-bcad1944-f21a-4c19-92d0-96ef2453d6e5 service nova] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.458478] env[70013]: DEBUG oslo_concurrency.lockutils [req-5023e738-303e-4607-92c6-453a219a98e8 req-bcad1944-f21a-4c19-92d0-96ef2453d6e5 service nova] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.458766] env[70013]: DEBUG nova.compute.manager [req-5023e738-303e-4607-92c6-453a219a98e8 req-bcad1944-f21a-4c19-92d0-96ef2453d6e5 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] No waiting events found dispatching network-vif-plugged-72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 927.459014] env[70013]: WARNING nova.compute.manager [req-5023e738-303e-4607-92c6-453a219a98e8 req-bcad1944-f21a-4c19-92d0-96ef2453d6e5 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Received unexpected event network-vif-plugged-72c3d938-1cba-4e01-8176-aa8e8a245c4c for instance with vm_state building and task_state spawning. [ 927.792145] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.794240] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.794482] env[70013]: DEBUG nova.network.neutron [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.839824] env[70013]: DEBUG nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 927.883798] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231047, 'name': Rename_Task, 'duration_secs': 0.175848} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.889401] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.889401] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dad336bc-6df8-466f-a6bf-478ce90269a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.896756] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 927.896756] env[70013]: value = "task-4231048" [ 927.896756] env[70013]: _type = "Task" [ 927.896756] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.926691] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231048, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.950947] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be32c6d-541a-48ea-89ba-0479c91f0cba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.961276] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a85752-c279-4638-bd24-10f0547098c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.003151] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fae8a55-3691-46bf-942b-1b785db94d16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.012310] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926b350d-abbb-4bee-8c9b-7144fb11bf5c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.030992] env[70013]: DEBUG nova.compute.provider_tree [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.034360] env[70013]: DEBUG nova.network.neutron [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Successfully created port: 3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 928.191134] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.191611] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.415698] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231048, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.540589] env[70013]: DEBUG nova.scheduler.client.report [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 928.616218] env[70013]: DEBUG nova.network.neutron [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 928.859610] env[70013]: DEBUG nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 928.864347] env[70013]: DEBUG nova.network.neutron [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updating instance_info_cache with network_info: [{"id": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "address": "fa:16:3e:aa:b4:a5", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72c3d938-1c", "ovs_interfaceid": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.899402] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 928.899745] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.899847] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 928.900011] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.900223] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 928.900437] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 928.900653] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 928.900749] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 928.900914] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 928.901090] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 928.901297] env[70013]: DEBUG nova.virt.hardware [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 928.902804] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c473225d-90f0-438f-af74-7135ddb1a949 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.920913] env[70013]: DEBUG oslo_vmware.api [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231048, 'name': PowerOnVM_Task, 'duration_secs': 0.548254} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.924183] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.924373] env[70013]: INFO nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Took 5.45 seconds to spawn the instance on the hypervisor. [ 928.924679] env[70013]: DEBUG nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 928.925615] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4f20e8-e2b3-4dc8-942b-97357f905af5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.931095] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b2131b-4833-4f6e-a1bc-f68a1a7f1f51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.051523] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.224s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.057959] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.220s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.062458] env[70013]: INFO nova.compute.claims [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.101076] env[70013]: INFO nova.scheduler.client.report [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Deleted allocations for instance a53389ce-d0c2-48ae-83e3-ee58fbdbfb01 [ 929.369550] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.370279] env[70013]: DEBUG nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Instance network_info: |[{"id": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "address": "fa:16:3e:aa:b4:a5", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72c3d938-1c", "ovs_interfaceid": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 929.370524] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:b4:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72c3d938-1cba-4e01-8176-aa8e8a245c4c', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.378659] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 929.378837] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 929.379086] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0484d25e-1f02-4b13-afe0-73843b19fec5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.401100] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.401100] env[70013]: value = "task-4231049" [ 929.401100] env[70013]: _type = "Task" [ 929.401100] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.410388] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231049, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.466789] env[70013]: INFO nova.compute.manager [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Took 49.37 seconds to build instance. [ 929.610841] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c38e8bc8-4350-448e-a026-f26bb28dbf7f tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "a53389ce-d0c2-48ae-83e3-ee58fbdbfb01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.387s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.792121] env[70013]: DEBUG nova.compute.manager [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Received event network-changed-72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 929.792572] env[70013]: DEBUG nova.compute.manager [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Refreshing instance network info cache due to event network-changed-72c3d938-1cba-4e01-8176-aa8e8a245c4c. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 929.792883] env[70013]: DEBUG oslo_concurrency.lockutils [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] Acquiring lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.792883] env[70013]: DEBUG oslo_concurrency.lockutils [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] Acquired lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.793142] env[70013]: DEBUG nova.network.neutron [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Refreshing network info cache for port 72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.915528] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231049, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.969394] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5aa28358-2241-4aa6-8874-0f88b8ebce5f tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "965480de-4fe2-48ce-94ad-0838d78239dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.050s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.146994] env[70013]: DEBUG nova.network.neutron [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Successfully updated port: 3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.416287] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231049, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.474673] env[70013]: DEBUG nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 930.651858] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.652041] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 930.652238] env[70013]: DEBUG nova.network.neutron [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.695654] env[70013]: DEBUG nova.network.neutron [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updated VIF entry in instance network info cache for port 72c3d938-1cba-4e01-8176-aa8e8a245c4c. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.696039] env[70013]: DEBUG nova.network.neutron [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updating instance_info_cache with network_info: [{"id": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "address": "fa:16:3e:aa:b4:a5", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72c3d938-1c", "ovs_interfaceid": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.728278] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352f73dc-2760-4328-9db0-cc59141ead8e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.739776] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb5bf3e-bbb5-441b-95c1-620dddb94f3a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.780672] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577b9e6d-4262-4647-aaf2-6ed58e80a6de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.786342] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "54d39dfd-6429-4009-8d26-22c1fa46243e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.786569] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.792784] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267ec934-dcbe-44a7-956c-7130704b7dcd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.807397] env[70013]: DEBUG nova.compute.provider_tree [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.914444] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231049, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.997337] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.177951] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.178474] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.194186] env[70013]: DEBUG nova.network.neutron [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 931.201512] env[70013]: DEBUG oslo_concurrency.lockutils [req-bd82b43e-0ecb-48ad-b8c2-f364fcd22f0f req-03751196-716e-427e-a342-166f6ca6c3b0 service nova] Releasing lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.310714] env[70013]: DEBUG nova.scheduler.client.report [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 931.366005] env[70013]: DEBUG nova.network.neutron [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fc25179-89", "ovs_interfaceid": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.415070] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231049, 'name': CreateVM_Task, 'duration_secs': 1.539583} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.415466] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.416034] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.416188] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.416509] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 931.416757] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-981766ca-6437-4301-b7d0-e21b994f937b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.422307] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 931.422307] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6e4df-31ee-c83f-7af4-71751e5b93dd" [ 931.422307] env[70013]: _type = "Task" [ 931.422307] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.430614] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6e4df-31ee-c83f-7af4-71751e5b93dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.821038] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.765s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.821038] env[70013]: DEBUG nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 931.823276] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.397s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.824739] env[70013]: INFO nova.compute.claims [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.869117] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.869482] env[70013]: DEBUG nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance network_info: |[{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fc25179-89", "ovs_interfaceid": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 931.870750] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:81:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3fc25179-89cf-42de-b6b6-f31806bcbce3', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.878275] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating folder: Project (0f5b0bd88f5046f098579c5d59e3f0c3). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 931.879202] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18edca8d-7fc1-4ad2-89c3-d03343bad3e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.893974] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created folder: Project (0f5b0bd88f5046f098579c5d59e3f0c3) in parent group-v836999. [ 931.894158] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating folder: Instances. Parent ref: group-v837144. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 931.894715] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa235190-426f-4385-b268-994c39e33d5a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.907741] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created folder: Instances in parent group-v837144. [ 931.908140] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 931.908687] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.908687] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14eba221-977f-4d06-94d1-204ae9838b44 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.936289] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6e4df-31ee-c83f-7af4-71751e5b93dd, 'name': SearchDatastore_Task, 'duration_secs': 0.037766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.937906] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.938189] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.938433] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.938584] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.938766] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.939040] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.939040] env[70013]: value = "task-4231052" [ 931.939040] env[70013]: _type = "Task" [ 931.939040] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.939231] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a8bae4c-a51f-4786-8276-57a55cba9347 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.967816] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231052, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.969427] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.969588] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.970304] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed7ef50d-c91a-417c-bd41-e1933cc88e2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.978414] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 931.978414] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528de0f9-84f3-37bf-ea7b-98a4558aedcd" [ 931.978414] env[70013]: _type = "Task" [ 931.978414] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.993022] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528de0f9-84f3-37bf-ea7b-98a4558aedcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.034538] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.034899] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.068473] env[70013]: DEBUG nova.compute.manager [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-vif-plugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 932.068694] env[70013]: DEBUG oslo_concurrency.lockutils [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.068902] env[70013]: DEBUG oslo_concurrency.lockutils [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.069171] env[70013]: DEBUG oslo_concurrency.lockutils [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.069454] env[70013]: DEBUG nova.compute.manager [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] No waiting events found dispatching network-vif-plugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 932.070070] env[70013]: WARNING nova.compute.manager [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received unexpected event network-vif-plugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 for instance with vm_state building and task_state spawning. [ 932.070070] env[70013]: DEBUG nova.compute.manager [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 932.070070] env[70013]: DEBUG nova.compute.manager [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing instance network info cache due to event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 932.070478] env[70013]: DEBUG oslo_concurrency.lockutils [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] Acquiring lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.070478] env[70013]: DEBUG oslo_concurrency.lockutils [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] Acquired lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.070478] env[70013]: DEBUG nova.network.neutron [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 932.092826] env[70013]: INFO nova.compute.manager [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Rebuilding instance [ 932.139895] env[70013]: DEBUG nova.compute.manager [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 932.141217] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aef8efe-4f8c-47b9-a301-78e296125071 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.331434] env[70013]: DEBUG nova.compute.utils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 932.335892] env[70013]: DEBUG nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 932.336156] env[70013]: DEBUG nova.network.neutron [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.388203] env[70013]: DEBUG nova.policy [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c4feb0b50fc4d83b93d6b9290848ee3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '946b799447b74a32b252c4f69fa900cd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 932.452388] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231052, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.488528] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528de0f9-84f3-37bf-ea7b-98a4558aedcd, 'name': SearchDatastore_Task, 'duration_secs': 0.014177} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.489678] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c116d20e-fe3d-4709-8981-a72f57692b59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.496427] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 932.496427] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52057f5e-58bf-f818-27d7-74140acc9d32" [ 932.496427] env[70013]: _type = "Task" [ 932.496427] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.505959] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52057f5e-58bf-f818-27d7-74140acc9d32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.793315] env[70013]: DEBUG nova.network.neutron [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Successfully created port: f6f11372-627a-4dc6-8fe9-e20fcc0a36ed {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.837032] env[70013]: DEBUG nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 932.916984] env[70013]: DEBUG nova.network.neutron [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updated VIF entry in instance network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.917252] env[70013]: DEBUG nova.network.neutron [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fc25179-89", "ovs_interfaceid": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.954564] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231052, 'name': CreateVM_Task, 'duration_secs': 0.645196} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.954886] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.955866] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.956190] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.956898] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 932.963020] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58216234-2a4b-443c-8e26-3c21c031e3bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.969313] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 932.969313] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522675f9-4d35-3a6e-c62b-f5533ea03399" [ 932.969313] env[70013]: _type = "Task" [ 932.969313] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.978423] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522675f9-4d35-3a6e-c62b-f5533ea03399, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.007065] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52057f5e-58bf-f818-27d7-74140acc9d32, 'name': SearchDatastore_Task, 'duration_secs': 0.01311} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.009821] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.010276] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 30fcb8c7-0947-426e-9aa5-081446e31c28/30fcb8c7-0947-426e-9aa5-081446e31c28.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.011581] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1684d0fb-8e98-490e-b5b0-3b1732f3b4ea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.018744] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 933.018744] env[70013]: value = "task-4231053" [ 933.018744] env[70013]: _type = "Task" [ 933.018744] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.031113] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231053, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.158475] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.159713] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-62602e29-d9e0-41c3-94c7-130bee2584f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.171027] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 933.171027] env[70013]: value = "task-4231054" [ 933.171027] env[70013]: _type = "Task" [ 933.171027] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.187565] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231054, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.338221] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f67227f-214c-4568-a2f5-c1690f517bbf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.355019] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead42bcb-a283-4054-97be-b43457605801 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.401101] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a01fd91-f58b-4fb8-a8d9-b7dfdc65bc4b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.412466] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b75e6c4-7f9c-4c2e-8733-4cca52b931da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.422028] env[70013]: DEBUG oslo_concurrency.lockutils [req-ceb0f790-7251-49a9-aeaa-b95077800221 req-20afca1b-9c9d-405a-bd18-6f310c45bfdd service nova] Releasing lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.430333] env[70013]: DEBUG nova.compute.provider_tree [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.480287] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522675f9-4d35-3a6e-c62b-f5533ea03399, 'name': SearchDatastore_Task, 'duration_secs': 0.031065} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.480752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.481021] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.481561] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.482091] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 933.482516] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.482930] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-081fa144-bac0-4c3c-a765-de09edf7cf62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.498359] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.498807] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 933.499494] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa5b2cff-fd7e-4584-bd37-ed413ff0748b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.507063] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 933.507063] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5295af40-db93-1e7e-4d85-cb313ed90aba" [ 933.507063] env[70013]: _type = "Task" [ 933.507063] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.519330] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5295af40-db93-1e7e-4d85-cb313ed90aba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.530492] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231053, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.705379] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231054, 'name': PowerOffVM_Task, 'duration_secs': 0.24777} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.705379] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.705379] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.705379] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12fb1fd-ef89-4bcb-a939-64ead712402d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.705379] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.705379] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ec902641-44cf-4aa9-b215-59a077b8c328 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.720851] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.722709] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.722829] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Deleting the datastore file [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.723101] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-246429b1-96f6-4411-9b8e-14ede9e9c3b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.734470] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 933.734470] env[70013]: value = "task-4231056" [ 933.734470] env[70013]: _type = "Task" [ 933.734470] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.748646] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.861502] env[70013]: DEBUG nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 933.888667] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 933.888909] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.889083] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 933.889274] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.889421] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 933.889565] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 933.889765] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 933.889917] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 933.890130] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 933.890373] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 933.890556] env[70013]: DEBUG nova.virt.hardware [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 933.892057] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeaa4976-854a-49eb-bbc7-34fdfa976e1f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.899729] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5557c7de-fe1b-4c91-930e-aeef22d54deb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.937744] env[70013]: DEBUG nova.scheduler.client.report [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 934.018327] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5295af40-db93-1e7e-4d85-cb313ed90aba, 'name': SearchDatastore_Task, 'duration_secs': 0.059261} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.019201] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16b602c7-3ab9-40c5-838b-a285923cb99c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.032311] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231053, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632645} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.033304] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 30fcb8c7-0947-426e-9aa5-081446e31c28/30fcb8c7-0947-426e-9aa5-081446e31c28.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.033525] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.033844] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 934.033844] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ece346-04de-8590-91b3-cd2e1b11e314" [ 934.033844] env[70013]: _type = "Task" [ 934.033844] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.034043] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18ff8654-f0de-4082-80f9-dbb1416a91a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.046257] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ece346-04de-8590-91b3-cd2e1b11e314, 'name': SearchDatastore_Task, 'duration_secs': 0.010411} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.049023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 934.049023] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 934.049023] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 934.049023] env[70013]: value = "task-4231057" [ 934.049023] env[70013]: _type = "Task" [ 934.049023] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.049023] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fd2624e-9d4e-4560-9f51-80d8c28931c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.061547] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231057, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.062262] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 934.062262] env[70013]: value = "task-4231058" [ 934.062262] env[70013]: _type = "Task" [ 934.062262] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.070523] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.153883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "67b90fec-a6fd-471d-a425-82bc63de572a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.154696] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "67b90fec-a6fd-471d-a425-82bc63de572a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.154696] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "67b90fec-a6fd-471d-a425-82bc63de572a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.154696] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "67b90fec-a6fd-471d-a425-82bc63de572a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.154696] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "67b90fec-a6fd-471d-a425-82bc63de572a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.157921] env[70013]: INFO nova.compute.manager [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Terminating instance [ 934.248670] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101329} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.248944] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.249154] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.249334] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.363991] env[70013]: DEBUG nova.compute.manager [req-26277f36-de46-4cd3-897e-b9fb088d9314 req-d271bab1-0b0a-4d30-a75d-ecf5366bec5a service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Received event network-vif-plugged-f6f11372-627a-4dc6-8fe9-e20fcc0a36ed {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 934.364310] env[70013]: DEBUG oslo_concurrency.lockutils [req-26277f36-de46-4cd3-897e-b9fb088d9314 req-d271bab1-0b0a-4d30-a75d-ecf5366bec5a service nova] Acquiring lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.364959] env[70013]: DEBUG oslo_concurrency.lockutils [req-26277f36-de46-4cd3-897e-b9fb088d9314 req-d271bab1-0b0a-4d30-a75d-ecf5366bec5a service nova] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.365186] env[70013]: DEBUG oslo_concurrency.lockutils [req-26277f36-de46-4cd3-897e-b9fb088d9314 req-d271bab1-0b0a-4d30-a75d-ecf5366bec5a service nova] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.365380] env[70013]: DEBUG nova.compute.manager [req-26277f36-de46-4cd3-897e-b9fb088d9314 req-d271bab1-0b0a-4d30-a75d-ecf5366bec5a service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] No waiting events found dispatching network-vif-plugged-f6f11372-627a-4dc6-8fe9-e20fcc0a36ed {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 934.365606] env[70013]: WARNING nova.compute.manager [req-26277f36-de46-4cd3-897e-b9fb088d9314 req-d271bab1-0b0a-4d30-a75d-ecf5366bec5a service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Received unexpected event network-vif-plugged-f6f11372-627a-4dc6-8fe9-e20fcc0a36ed for instance with vm_state building and task_state spawning. [ 934.445376] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.448786] env[70013]: DEBUG nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 934.449684] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.342s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.451701] env[70013]: INFO nova.compute.claims [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.519607] env[70013]: DEBUG nova.network.neutron [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Successfully updated port: f6f11372-627a-4dc6-8fe9-e20fcc0a36ed {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.561801] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231057, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078372} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.563026] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.563927] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837d70fd-964c-45bb-a77c-188c24849bb8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.597385] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 30fcb8c7-0947-426e-9aa5-081446e31c28/30fcb8c7-0947-426e-9aa5-081446e31c28.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.597385] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231058, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474161} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.597385] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1868a42-1bea-437f-a35d-16a4acc2243f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.613613] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.613613] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.613613] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06daa8a4-fd04-4b73-96a9-db76ef87b1fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.620496] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 934.620496] env[70013]: value = "task-4231059" [ 934.620496] env[70013]: _type = "Task" [ 934.620496] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.622033] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 934.622033] env[70013]: value = "task-4231060" [ 934.622033] env[70013]: _type = "Task" [ 934.622033] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.635198] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231060, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.639219] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231059, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.661910] env[70013]: DEBUG nova.compute.manager [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 934.661910] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.662645] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3808183-7e28-431e-9295-47c9cc676cee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.674180] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.674180] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01781cf6-8fe6-4682-b09a-9c59151ada6e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.679190] env[70013]: DEBUG oslo_vmware.api [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 934.679190] env[70013]: value = "task-4231061" [ 934.679190] env[70013]: _type = "Task" [ 934.679190] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.688263] env[70013]: DEBUG oslo_vmware.api [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4231061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.957144] env[70013]: DEBUG nova.compute.utils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 934.962496] env[70013]: DEBUG nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 934.962496] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 935.026092] env[70013]: DEBUG nova.policy [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '804d90ab34784afe864a6b76379619bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4eef90bcd1794b788b1340d14c6df552', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 935.026092] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "refresh_cache-626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.026092] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "refresh_cache-626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 935.026390] env[70013]: DEBUG nova.network.neutron [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.136054] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231060, 'name': ReconfigVM_Task, 'duration_secs': 0.306873} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.139566] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 30fcb8c7-0947-426e-9aa5-081446e31c28/30fcb8c7-0947-426e-9aa5-081446e31c28.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.140401] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231059, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073466} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.140669] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4379a7c1-c2a0-4a14-a508-d75e8a07fb78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.142561] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.143208] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d63e11e-1807-4ad5-a853-082f43156e52 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.171862] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.174107] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6371754e-73c0-4435-91a4-23d62d66ce07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.190477] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 935.190477] env[70013]: value = "task-4231062" [ 935.190477] env[70013]: _type = "Task" [ 935.190477] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.200189] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 935.200189] env[70013]: value = "task-4231063" [ 935.200189] env[70013]: _type = "Task" [ 935.200189] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.206387] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231062, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.206645] env[70013]: DEBUG oslo_vmware.api [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4231061, 'name': PowerOffVM_Task, 'duration_secs': 0.242205} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.210187] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.210405] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.210673] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c130acfb-1b6a-4cc3-8b55-05cdd596b00d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.219524] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231063, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.289902] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.289902] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.290127] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Deleting the datastore file [datastore1] 67b90fec-a6fd-471d-a425-82bc63de572a {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.292617] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 935.292953] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.293071] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 935.293268] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.293423] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 935.293567] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 935.294036] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 935.296032] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 935.296032] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 935.296032] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 935.296032] env[70013]: DEBUG nova.virt.hardware [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 935.296032] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7db5acf4-bb6e-4132-9fe0-9cc89d01d75a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.299021] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300db79e-80e4-4c9f-8bb8-76c88930f751 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.314111] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8693797c-883f-4587-bf7a-c3cf9252eacc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.318164] env[70013]: DEBUG oslo_vmware.api [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for the task: (returnval){ [ 935.318164] env[70013]: value = "task-4231065" [ 935.318164] env[70013]: _type = "Task" [ 935.318164] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.331732] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.338165] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 935.339190] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 935.339550] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5735655-525d-4ce2-a54b-fa5d12d97fef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.357019] env[70013]: DEBUG oslo_vmware.api [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4231065, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.364019] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.364019] env[70013]: value = "task-4231066" [ 935.364019] env[70013]: _type = "Task" [ 935.364019] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.371239] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Successfully created port: 6114ded0-69f6-4591-bc0f-defc91e44362 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.377235] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231066, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.461748] env[70013]: DEBUG nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 935.579145] env[70013]: DEBUG nova.network.neutron [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.708628] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231062, 'name': Rename_Task, 'duration_secs': 0.307805} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.713137] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.713704] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c05e907f-447a-4c9e-83b7-311987da0dfc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.724705] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231063, 'name': ReconfigVM_Task, 'duration_secs': 0.333576} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.727542] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.727680] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 935.727680] env[70013]: value = "task-4231067" [ 935.727680] env[70013]: _type = "Task" [ 935.727680] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.727884] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1546f46-dafb-468d-a9d7-df710edef325 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.739275] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.743584] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 935.743584] env[70013]: value = "task-4231068" [ 935.743584] env[70013]: _type = "Task" [ 935.743584] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.754267] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231068, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.832927] env[70013]: DEBUG oslo_vmware.api [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Task: {'id': task-4231065, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254733} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.833231] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.833428] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.833621] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.833805] env[70013]: INFO nova.compute.manager [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Took 1.17 seconds to destroy the instance on the hypervisor. [ 935.834073] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 935.834278] env[70013]: DEBUG nova.compute.manager [-] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 935.834377] env[70013]: DEBUG nova.network.neutron [-] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.847985] env[70013]: DEBUG nova.network.neutron [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Updating instance_info_cache with network_info: [{"id": "f6f11372-627a-4dc6-8fe9-e20fcc0a36ed", "address": "fa:16:3e:98:c1:c5", "network": {"id": "0f803fec-8c1a-495e-b561-9ae7e3b7c4e2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-731158857-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "946b799447b74a32b252c4f69fa900cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6f11372-62", "ovs_interfaceid": "f6f11372-627a-4dc6-8fe9-e20fcc0a36ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.881338] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231066, 'name': CreateVM_Task, 'duration_secs': 0.419731} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.882249] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Successfully created port: 0dd5389a-041f-43b0-949a-78f2655dc2ee {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.883645] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 935.886032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.886032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 935.886032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 935.886032] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82635bbe-919e-48e4-9fe6-c98f7f949a24 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.893667] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 935.893667] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526f1b8a-381e-b500-5a96-0d7ca0ae7958" [ 935.893667] env[70013]: _type = "Task" [ 935.893667] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.905543] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526f1b8a-381e-b500-5a96-0d7ca0ae7958, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.062492] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffea1edc-e826-4ea0-a193-e51915b306cd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.072092] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b3ce39-ddab-4609-b42b-dec6a7e17f73 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.111431] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8fdd4e-7eb9-440f-bf6a-419401d677b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.120146] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b51842-156a-470d-9ccd-6f19b55dddf2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.135826] env[70013]: DEBUG nova.compute.provider_tree [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.182169] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Successfully created port: 894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.242993] env[70013]: DEBUG oslo_vmware.api [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231067, 'name': PowerOnVM_Task, 'duration_secs': 0.485435} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.243541] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.243915] env[70013]: INFO nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Took 10.07 seconds to spawn the instance on the hypervisor. [ 936.244225] env[70013]: DEBUG nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 936.245241] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd85bc69-fcd9-431b-9f33-28c6eedf9143 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.264015] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231068, 'name': Rename_Task, 'duration_secs': 0.28117} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.266340] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.266340] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25211b13-b2d5-4934-8fb9-8604766b1eb7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.275906] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 936.275906] env[70013]: value = "task-4231069" [ 936.275906] env[70013]: _type = "Task" [ 936.275906] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.288043] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.355017] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "refresh_cache-626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 936.355017] env[70013]: DEBUG nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Instance network_info: |[{"id": "f6f11372-627a-4dc6-8fe9-e20fcc0a36ed", "address": "fa:16:3e:98:c1:c5", "network": {"id": "0f803fec-8c1a-495e-b561-9ae7e3b7c4e2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-731158857-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "946b799447b74a32b252c4f69fa900cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6f11372-62", "ovs_interfaceid": "f6f11372-627a-4dc6-8fe9-e20fcc0a36ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 936.355017] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:c1:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a55f45a-d631-4ebc-b73b-8a30bd0a32a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6f11372-627a-4dc6-8fe9-e20fcc0a36ed', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.361647] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Creating folder: Project (946b799447b74a32b252c4f69fa900cd). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 936.362100] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-57a1ccfa-1049-4475-9638-e596b10c3cee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.375340] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Created folder: Project (946b799447b74a32b252c4f69fa900cd) in parent group-v836999. [ 936.375851] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Creating folder: Instances. Parent ref: group-v837148. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 936.376286] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2a0d190-609d-4857-991c-f9484851d034 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.388694] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Created folder: Instances in parent group-v837148. [ 936.388694] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 936.388694] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.388694] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ce01e49-e8f0-41fd-b1d1-b1f8c1e1c500 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.415376] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526f1b8a-381e-b500-5a96-0d7ca0ae7958, 'name': SearchDatastore_Task, 'duration_secs': 0.011329} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.416826] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 936.417143] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.417897] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.418162] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 936.418325] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.418625] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.418625] env[70013]: value = "task-4231072" [ 936.418625] env[70013]: _type = "Task" [ 936.418625] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.420077] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64aabb27-3158-49e8-b92f-5dca804654e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.423908] env[70013]: DEBUG nova.compute.manager [req-1cba26ef-f3ba-4d1f-bcf8-b76e483e8503 req-1bf7392d-a6b6-4452-b623-177e0c6b7106 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Received event network-vif-deleted-9e2b8cbf-ea68-4961-ae17-64f6753e3974 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 936.424152] env[70013]: INFO nova.compute.manager [req-1cba26ef-f3ba-4d1f-bcf8-b76e483e8503 req-1bf7392d-a6b6-4452-b623-177e0c6b7106 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Neutron deleted interface 9e2b8cbf-ea68-4961-ae17-64f6753e3974; detaching it from the instance and deleting it from the info cache [ 936.424380] env[70013]: DEBUG nova.network.neutron [req-1cba26ef-f3ba-4d1f-bcf8-b76e483e8503 req-1bf7392d-a6b6-4452-b623-177e0c6b7106 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.437480] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231072, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.438986] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.443236] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.443236] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d940f12-0abd-4249-9cdd-d71726d058db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.445959] env[70013]: DEBUG nova.compute.manager [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Received event network-changed-f6f11372-627a-4dc6-8fe9-e20fcc0a36ed {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 936.446212] env[70013]: DEBUG nova.compute.manager [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Refreshing instance network info cache due to event network-changed-f6f11372-627a-4dc6-8fe9-e20fcc0a36ed. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 936.446491] env[70013]: DEBUG oslo_concurrency.lockutils [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] Acquiring lock "refresh_cache-626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.446674] env[70013]: DEBUG oslo_concurrency.lockutils [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] Acquired lock "refresh_cache-626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 936.446892] env[70013]: DEBUG nova.network.neutron [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Refreshing network info cache for port f6f11372-627a-4dc6-8fe9-e20fcc0a36ed {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.453836] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 936.453836] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f893d6-6e4d-c122-1412-6cf2a7e55be7" [ 936.453836] env[70013]: _type = "Task" [ 936.453836] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.465599] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f893d6-6e4d-c122-1412-6cf2a7e55be7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.476559] env[70013]: DEBUG nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 936.514604] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 936.514919] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.515017] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 936.516423] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.516685] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 936.516853] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 936.517222] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 936.517833] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 936.518078] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 936.519202] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 936.519407] env[70013]: DEBUG nova.virt.hardware [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 936.521229] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbad2e2-78e5-4a0d-9cc5-b865257adb35 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.533222] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ff5be8-8d2f-46f9-ad87-b00821124ff8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.639015] env[70013]: DEBUG nova.scheduler.client.report [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 936.781499] env[70013]: INFO nova.compute.manager [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Took 55.49 seconds to build instance. [ 936.789587] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231069, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.808707] env[70013]: DEBUG nova.network.neutron [-] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.933609] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e21daf27-9d41-4ae4-a49e-3650a9f6cbcb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.948856] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231072, 'name': CreateVM_Task, 'duration_secs': 0.429315} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.953299] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.954701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.955483] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 936.956236] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 936.960025] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576ce08c-2f7d-4ead-9ffa-ae23f94c4e20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.972059] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67832a8d-d4c5-4fac-8e56-78ceaed7406e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.982685] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 936.982685] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52675a11-b361-b91c-d390-76f88a44a5e8" [ 936.982685] env[70013]: _type = "Task" [ 936.982685] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.987023] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f893d6-6e4d-c122-1412-6cf2a7e55be7, 'name': SearchDatastore_Task, 'duration_secs': 0.01723} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.000652] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f6f7d34-4e85-4245-89d3-87e009f03a4b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.003375] env[70013]: DEBUG nova.compute.manager [req-1cba26ef-f3ba-4d1f-bcf8-b76e483e8503 req-1bf7392d-a6b6-4452-b623-177e0c6b7106 service nova] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Detach interface failed, port_id=9e2b8cbf-ea68-4961-ae17-64f6753e3974, reason: Instance 67b90fec-a6fd-471d-a425-82bc63de572a could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 937.007607] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 937.007607] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e14d2-2a78-39fc-0587-8fa81adba9bf" [ 937.007607] env[70013]: _type = "Task" [ 937.007607] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.010825] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52675a11-b361-b91c-d390-76f88a44a5e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.019621] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e14d2-2a78-39fc-0587-8fa81adba9bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.144077] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.694s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.144677] env[70013]: DEBUG nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 937.147704] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 32.490s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.147890] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.148054] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 937.148392] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.119s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.148597] env[70013]: DEBUG nova.objects.instance [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lazy-loading 'resources' on Instance uuid 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.150890] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96241f88-8faf-4b2a-8913-7236e097b65e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.160114] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fa0d2f-f77f-4cd3-a909-ffbdb2be8d85 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.182642] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a237ab8-6202-4a0d-b4c2-e96ebf8754a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.192192] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb2017e-97af-4e6c-9a96-0f324b9acf00 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.228855] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179631MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 937.229023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.288107] env[70013]: DEBUG oslo_concurrency.lockutils [None req-52517634-fdf2-494c-8359-52b97cb4080e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.496s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.288478] env[70013]: DEBUG oslo_vmware.api [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231069, 'name': PowerOnVM_Task, 'duration_secs': 0.667913} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.289743] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.289953] env[70013]: INFO nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Took 8.43 seconds to spawn the instance on the hypervisor. [ 937.290154] env[70013]: DEBUG nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 937.291354] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f554e9-e48b-4584-b3e2-ce47a3e3c110 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.313634] env[70013]: INFO nova.compute.manager [-] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Took 1.48 seconds to deallocate network for instance. [ 937.451548] env[70013]: DEBUG nova.network.neutron [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Updated VIF entry in instance network info cache for port f6f11372-627a-4dc6-8fe9-e20fcc0a36ed. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.451918] env[70013]: DEBUG nova.network.neutron [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Updating instance_info_cache with network_info: [{"id": "f6f11372-627a-4dc6-8fe9-e20fcc0a36ed", "address": "fa:16:3e:98:c1:c5", "network": {"id": "0f803fec-8c1a-495e-b561-9ae7e3b7c4e2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-731158857-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "946b799447b74a32b252c4f69fa900cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6f11372-62", "ovs_interfaceid": "f6f11372-627a-4dc6-8fe9-e20fcc0a36ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.499856] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52675a11-b361-b91c-d390-76f88a44a5e8, 'name': SearchDatastore_Task, 'duration_secs': 0.034449} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.500409] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 937.500409] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.500554] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.521342] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e14d2-2a78-39fc-0587-8fa81adba9bf, 'name': SearchDatastore_Task, 'duration_secs': 0.037793} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.521629] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 937.521887] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.522215] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 937.522404] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.522621] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c20ecc5b-7395-4468-92c7-9398e903d6bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.525620] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12c6bdb4-70f6-40f4-9970-420b203c605a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.534827] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 937.534827] env[70013]: value = "task-4231073" [ 937.534827] env[70013]: _type = "Task" [ 937.534827] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.539908] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.541018] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 937.541304] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42dd8e6f-1836-4e2b-ad46-200a476c5ef1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.547162] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.552526] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 937.552526] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5271b882-9c7a-cb03-1a70-99abc7079213" [ 937.552526] env[70013]: _type = "Task" [ 937.552526] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.562099] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5271b882-9c7a-cb03-1a70-99abc7079213, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.652465] env[70013]: DEBUG nova.compute.utils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 937.653994] env[70013]: DEBUG nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 937.654256] env[70013]: DEBUG nova.network.neutron [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 937.763229] env[70013]: DEBUG nova.policy [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 937.791526] env[70013]: DEBUG nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 937.813757] env[70013]: INFO nova.compute.manager [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Took 47.50 seconds to build instance. [ 937.823265] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.866375] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Successfully updated port: 6114ded0-69f6-4591-bc0f-defc91e44362 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.955463] env[70013]: DEBUG oslo_concurrency.lockutils [req-c739b11c-40d4-411c-aaf7-d331af3788f9 req-9fd53743-4621-4598-a56e-a77ab3cc0ed2 service nova] Releasing lock "refresh_cache-626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 938.049922] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231073, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.062466] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5271b882-9c7a-cb03-1a70-99abc7079213, 'name': SearchDatastore_Task, 'duration_secs': 0.015099} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.066259] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a98ce94-4eab-47a1-bc5d-53730feb00c6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.072898] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 938.072898] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b5794a-a40b-9623-f461-483958fc096c" [ 938.072898] env[70013]: _type = "Task" [ 938.072898] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.085787] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b5794a-a40b-9623-f461-483958fc096c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.164246] env[70013]: DEBUG nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 938.233200] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4eb6b1c-22d4-480f-9d4e-11f940e74ffc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.242959] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ec934e-2549-46c3-96d5-b16e5c7bad80 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.284349] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "30fcb8c7-0947-426e-9aa5-081446e31c28" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.284522] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.284699] env[70013]: INFO nova.compute.manager [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Shelving [ 938.288035] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54217dc-74b4-4f10-ac99-9a2ca9af407c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.299866] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdde9ed-2588-4206-9308-1f75958eaae8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.319780] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a9f98d2c-ff86-459e-b5b0-09a322e77e2b tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.294s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.320376] env[70013]: DEBUG nova.compute.provider_tree [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.324970] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.546496] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231073, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.697379} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.546894] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.547607] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.547970] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a546ba4f-1527-41d8-b3ef-e8fb84784f20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.558217] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 938.558217] env[70013]: value = "task-4231074" [ 938.558217] env[70013]: _type = "Task" [ 938.558217] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.569090] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231074, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.584079] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b5794a-a40b-9623-f461-483958fc096c, 'name': SearchDatastore_Task, 'duration_secs': 0.05432} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.584507] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 938.584696] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 938.585042] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ccfe7e2-c414-41e0-9cb4-8afaafdac714 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.594874] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 938.594874] env[70013]: value = "task-4231075" [ 938.594874] env[70013]: _type = "Task" [ 938.594874] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.606052] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.626119] env[70013]: DEBUG nova.network.neutron [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Successfully created port: 3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.736640] env[70013]: DEBUG nova.compute.manager [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-vif-plugged-6114ded0-69f6-4591-bc0f-defc91e44362 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 938.737163] env[70013]: DEBUG oslo_concurrency.lockutils [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] Acquiring lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.737382] env[70013]: DEBUG oslo_concurrency.lockutils [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.737850] env[70013]: DEBUG oslo_concurrency.lockutils [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.737941] env[70013]: DEBUG nova.compute.manager [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] No waiting events found dispatching network-vif-plugged-6114ded0-69f6-4591-bc0f-defc91e44362 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 938.738255] env[70013]: WARNING nova.compute.manager [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received unexpected event network-vif-plugged-6114ded0-69f6-4591-bc0f-defc91e44362 for instance with vm_state building and task_state spawning. [ 938.738538] env[70013]: DEBUG nova.compute.manager [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-changed-6114ded0-69f6-4591-bc0f-defc91e44362 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 938.738825] env[70013]: DEBUG nova.compute.manager [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Refreshing instance network info cache due to event network-changed-6114ded0-69f6-4591-bc0f-defc91e44362. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 938.739180] env[70013]: DEBUG oslo_concurrency.lockutils [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] Acquiring lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.739428] env[70013]: DEBUG oslo_concurrency.lockutils [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] Acquired lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 938.739741] env[70013]: DEBUG nova.network.neutron [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Refreshing network info cache for port 6114ded0-69f6-4591-bc0f-defc91e44362 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.828199] env[70013]: DEBUG nova.scheduler.client.report [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 938.830171] env[70013]: DEBUG nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 939.069540] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231074, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080667} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.069890] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.071400] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74a0689-de4d-42ab-b383-3844542335a3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.095182] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.095593] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6416c42d-cbc4-4056-942a-5363ea4ec024 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.121454] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49861} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.122826] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 939.123077] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.123392] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 939.123392] env[70013]: value = "task-4231076" [ 939.123392] env[70013]: _type = "Task" [ 939.123392] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.123586] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfaed89d-cc99-47d9-bc96-db5a7f5f5cf9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.140679] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231076, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.142015] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 939.142015] env[70013]: value = "task-4231077" [ 939.142015] env[70013]: _type = "Task" [ 939.142015] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.178918] env[70013]: DEBUG nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 939.208207] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 939.208537] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.209831] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 939.209831] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.209831] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 939.209831] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 939.209831] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 939.209831] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 939.210180] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 939.210180] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 939.210394] env[70013]: DEBUG nova.virt.hardware [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 939.212923] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3517171-f209-40b0-a5f7-71e594e7a190 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.223074] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb31722-50eb-4f7e-b2db-c77389023d11 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.286660] env[70013]: DEBUG nova.network.neutron [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 939.301901] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.304456] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed7624c9-6145-428d-b220-47685b165fa9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.311453] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 939.311453] env[70013]: value = "task-4231078" [ 939.311453] env[70013]: _type = "Task" [ 939.311453] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.322621] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.338694] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.190s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.344040] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.707s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.345748] env[70013]: INFO nova.compute.claims [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.371488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.377376] env[70013]: INFO nova.scheduler.client.report [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Deleted allocations for instance 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f [ 939.400370] env[70013]: DEBUG nova.network.neutron [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.638482] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231076, 'name': ReconfigVM_Task, 'duration_secs': 0.416291} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.639549] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd/965480de-4fe2-48ce-94ad-0838d78239dd.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.640268] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbe904b5-3935-46df-adba-8a492abe8b57 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.650028] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 939.650028] env[70013]: value = "task-4231079" [ 939.650028] env[70013]: _type = "Task" [ 939.650028] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.657855] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100541} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.658905] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.667258] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7424122-ca4b-4fd8-8f13-1a467f8305db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.674741] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231079, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.696608] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.696978] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60c3df48-f4ca-4c89-8704-93f7d229933b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.723527] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 939.723527] env[70013]: value = "task-4231080" [ 939.723527] env[70013]: _type = "Task" [ 939.723527] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.733015] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.822759] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231078, 'name': PowerOffVM_Task, 'duration_secs': 0.219077} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.823347] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.824408] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7799c71f-3753-4d75-bb6f-6efd7f0571f2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.848683] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78784fa-05c0-405a-950c-bb4d326f74a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.890991] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85b0d15c-321e-423a-bd0c-cc3265bcbefe tempest-ServerAddressesTestJSON-1300084773 tempest-ServerAddressesTestJSON-1300084773-project-member] Lock "0a56e30f-fb07-4a5e-9d69-9603fd38ff9f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.647s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.902797] env[70013]: DEBUG oslo_concurrency.lockutils [req-c928847e-f1b0-40c8-a9a5-b0ebc937b291 req-73f61292-4beb-46cb-9c52-529253614053 service nova] Releasing lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.055188] env[70013]: DEBUG nova.compute.manager [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 940.055397] env[70013]: DEBUG nova.compute.manager [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing instance network info cache due to event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 940.055657] env[70013]: DEBUG oslo_concurrency.lockutils [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] Acquiring lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.055784] env[70013]: DEBUG oslo_concurrency.lockutils [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] Acquired lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 940.055940] env[70013]: DEBUG nova.network.neutron [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.165932] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231079, 'name': Rename_Task, 'duration_secs': 0.160165} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.166907] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.167596] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc6a5ae5-17c6-40c3-8e1a-a3a4a35495d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.180017] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 940.180017] env[70013]: value = "task-4231081" [ 940.180017] env[70013]: _type = "Task" [ 940.180017] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.194906] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.238025] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.361705] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Successfully updated port: 0dd5389a-041f-43b0-949a-78f2655dc2ee {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.364811] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 940.365167] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d3cd3d60-bac5-4062-a6bd-aaa20e8090f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.381229] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 940.381229] env[70013]: value = "task-4231082" [ 940.381229] env[70013]: _type = "Task" [ 940.381229] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.394864] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231082, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.694020] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231081, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.734156] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231080, 'name': ReconfigVM_Task, 'duration_secs': 0.893843} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.736723] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.737656] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-387f613d-a327-4b6f-b6af-1d367653bae3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.746331] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 940.746331] env[70013]: value = "task-4231083" [ 940.746331] env[70013]: _type = "Task" [ 940.746331] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.758768] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231083, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.896806] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231082, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.916049] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b273e73-873f-48c5-9fd3-c0f671170052 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.924192] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb55f47-46ae-440e-8984-1f2f3258c473 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.967384] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6750b8e4-f9f9-47b0-b699-7a7eebcdd8da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.976532] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccd2303-85a4-4be4-ba11-08833c435b83 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.987028] env[70013]: DEBUG nova.compute.manager [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-vif-plugged-0dd5389a-041f-43b0-949a-78f2655dc2ee {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 940.987028] env[70013]: DEBUG oslo_concurrency.lockutils [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] Acquiring lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.987028] env[70013]: DEBUG oslo_concurrency.lockutils [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.987028] env[70013]: DEBUG oslo_concurrency.lockutils [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.987028] env[70013]: DEBUG nova.compute.manager [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] No waiting events found dispatching network-vif-plugged-0dd5389a-041f-43b0-949a-78f2655dc2ee {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 940.987028] env[70013]: WARNING nova.compute.manager [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received unexpected event network-vif-plugged-0dd5389a-041f-43b0-949a-78f2655dc2ee for instance with vm_state building and task_state spawning. [ 940.987028] env[70013]: DEBUG nova.compute.manager [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-changed-0dd5389a-041f-43b0-949a-78f2655dc2ee {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 940.987028] env[70013]: DEBUG nova.compute.manager [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Refreshing instance network info cache due to event network-changed-0dd5389a-041f-43b0-949a-78f2655dc2ee. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 940.987028] env[70013]: DEBUG oslo_concurrency.lockutils [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] Acquiring lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.987028] env[70013]: DEBUG oslo_concurrency.lockutils [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] Acquired lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 940.987028] env[70013]: DEBUG nova.network.neutron [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Refreshing network info cache for port 0dd5389a-041f-43b0-949a-78f2655dc2ee {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.997330] env[70013]: DEBUG nova.compute.provider_tree [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 941.197208] env[70013]: DEBUG oslo_vmware.api [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231081, 'name': PowerOnVM_Task, 'duration_secs': 0.525379} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.197587] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.197732] env[70013]: DEBUG nova.compute.manager [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.198593] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b33ac8-a795-472e-a0e4-5b10d4d08520 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.260673] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231083, 'name': Rename_Task, 'duration_secs': 0.235047} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.260984] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.261463] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39857fc4-12b6-4c62-94ea-590acd00f964 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.269741] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 941.269741] env[70013]: value = "task-4231084" [ 941.269741] env[70013]: _type = "Task" [ 941.269741] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.282850] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231084, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.321539] env[70013]: DEBUG nova.network.neutron [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Successfully updated port: 3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.394784] env[70013]: DEBUG nova.network.neutron [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updated VIF entry in instance network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.395179] env[70013]: DEBUG nova.network.neutron [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fc25179-89", "ovs_interfaceid": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.405922] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231082, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.529292] env[70013]: ERROR nova.scheduler.client.report [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [req-87e0233f-285c-476c-9d7e-5168c5079cf0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-87e0233f-285c-476c-9d7e-5168c5079cf0"}]} [ 941.529292] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.186s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.529794] env[70013]: ERROR nova.compute.manager [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Traceback (most recent call last): [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] yield [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] self.set_inventory_for_provider( [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-87e0233f-285c-476c-9d7e-5168c5079cf0"}]} [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] During handling of the above exception, another exception occurred: [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Traceback (most recent call last): [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] with self.rt.instance_claim(context, instance, node, allocs, [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] return f(*args, **kwargs) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] self._update(elevated, cn) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] self._update_to_placement(context, compute_node, startup) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] return attempt.get(self._wrap_exception) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] raise value [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] self.reportclient.update_from_provider_tree( [ 941.529794] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 941.530752] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] with catch_all(pd.uuid): [ 941.530752] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 941.530752] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] self.gen.throw(typ, value, traceback) [ 941.530752] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 941.530752] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] raise exception.ResourceProviderSyncFailed() [ 941.530752] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 941.530752] env[70013]: ERROR nova.compute.manager [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] [ 941.530752] env[70013]: DEBUG nova.compute.utils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 941.531830] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.799s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.533800] env[70013]: INFO nova.compute.claims [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.537160] env[70013]: DEBUG nova.compute.manager [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Build of instance 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 941.537578] env[70013]: DEBUG nova.compute.manager [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 941.537813] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Acquiring lock "refresh_cache-55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.537985] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Acquired lock "refresh_cache-55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.538139] env[70013]: DEBUG nova.network.neutron [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.580413] env[70013]: DEBUG nova.network.neutron [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.724793] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.744162] env[70013]: DEBUG nova.network.neutron [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.791631] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231084, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.825463] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.825892] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.826114] env[70013]: DEBUG nova.network.neutron [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.897910] env[70013]: DEBUG oslo_concurrency.lockutils [req-89991467-4396-4a0e-8fd1-3fbc40a82244 req-5c08fda1-1c08-4041-b82e-b664e5dae694 service nova] Releasing lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.898475] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231082, 'name': CreateSnapshot_Task, 'duration_secs': 1.063376} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.898713] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 941.899554] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5953ebe6-c1ef-4bc2-80d5-98ab31f3458e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.070484] env[70013]: DEBUG nova.network.neutron [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.201981] env[70013]: DEBUG nova.network.neutron [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.238323] env[70013]: DEBUG nova.compute.manager [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-vif-plugged-3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 942.238323] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.238323] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.238323] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.238323] env[70013]: DEBUG nova.compute.manager [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] No waiting events found dispatching network-vif-plugged-3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 942.238323] env[70013]: WARNING nova.compute.manager [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received unexpected event network-vif-plugged-3e38f8cd-84a0-48be-9215-3dec9c9ced94 for instance with vm_state building and task_state spawning. [ 942.239092] env[70013]: DEBUG nova.compute.manager [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-changed-3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 942.239412] env[70013]: DEBUG nova.compute.manager [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing instance network info cache due to event network-changed-3e38f8cd-84a0-48be-9215-3dec9c9ced94. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 942.239739] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.247313] env[70013]: DEBUG oslo_concurrency.lockutils [req-b70d38f9-b107-4c83-bf3c-ae628c7d894d req-7b80b843-8947-41ae-81b6-5fb534c9d0d9 service nova] Releasing lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.283441] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.283774] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.284140] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.284485] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.284651] env[70013]: INFO nova.compute.manager [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Rebooting instance [ 942.294608] env[70013]: DEBUG oslo_vmware.api [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231084, 'name': PowerOnVM_Task, 'duration_secs': 0.568908} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.295550] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 942.295803] env[70013]: INFO nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Took 8.43 seconds to spawn the instance on the hypervisor. [ 942.296037] env[70013]: DEBUG nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 942.297257] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a02732a-00a3-4903-80cb-32921648db24 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.396417] env[70013]: DEBUG nova.network.neutron [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.419190] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 942.422644] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f6c790e9-3678-47c4-b3c3-306bebfc4f43 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.433401] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 942.433401] env[70013]: value = "task-4231085" [ 942.433401] env[70013]: _type = "Task" [ 942.433401] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.444670] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231085, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.568923] env[70013]: DEBUG nova.scheduler.client.report [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 942.591760] env[70013]: DEBUG nova.scheduler.client.report [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 942.592085] env[70013]: DEBUG nova.compute.provider_tree [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 942.610962] env[70013]: DEBUG nova.scheduler.client.report [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 942.632980] env[70013]: DEBUG nova.scheduler.client.report [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 942.644337] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "965480de-4fe2-48ce-94ad-0838d78239dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.644614] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "965480de-4fe2-48ce-94ad-0838d78239dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.644817] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "965480de-4fe2-48ce-94ad-0838d78239dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.644999] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "965480de-4fe2-48ce-94ad-0838d78239dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.645187] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "965480de-4fe2-48ce-94ad-0838d78239dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.648948] env[70013]: INFO nova.compute.manager [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Terminating instance [ 942.658394] env[70013]: DEBUG nova.network.neutron [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.705426] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Releasing lock "refresh_cache-55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.705700] env[70013]: DEBUG nova.compute.manager [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 942.707744] env[70013]: DEBUG nova.compute.manager [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 942.707744] env[70013]: DEBUG nova.network.neutron [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.737409] env[70013]: DEBUG nova.network.neutron [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.809846] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.809846] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.809846] env[70013]: DEBUG nova.network.neutron [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.829530] env[70013]: INFO nova.compute.manager [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Took 43.01 seconds to build instance. [ 942.952403] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231085, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.021077] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Successfully updated port: 894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.023960] env[70013]: DEBUG nova.compute.manager [req-fe566082-a725-4519-b81e-c92b57d8af8d req-97a1b5c5-7dde-4f63-90b3-015c1ed31739 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-vif-plugged-894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 943.024228] env[70013]: DEBUG oslo_concurrency.lockutils [req-fe566082-a725-4519-b81e-c92b57d8af8d req-97a1b5c5-7dde-4f63-90b3-015c1ed31739 service nova] Acquiring lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.024725] env[70013]: DEBUG oslo_concurrency.lockutils [req-fe566082-a725-4519-b81e-c92b57d8af8d req-97a1b5c5-7dde-4f63-90b3-015c1ed31739 service nova] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.024725] env[70013]: DEBUG oslo_concurrency.lockutils [req-fe566082-a725-4519-b81e-c92b57d8af8d req-97a1b5c5-7dde-4f63-90b3-015c1ed31739 service nova] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.024816] env[70013]: DEBUG nova.compute.manager [req-fe566082-a725-4519-b81e-c92b57d8af8d req-97a1b5c5-7dde-4f63-90b3-015c1ed31739 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] No waiting events found dispatching network-vif-plugged-894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 943.024940] env[70013]: WARNING nova.compute.manager [req-fe566082-a725-4519-b81e-c92b57d8af8d req-97a1b5c5-7dde-4f63-90b3-015c1ed31739 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received unexpected event network-vif-plugged-894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a for instance with vm_state building and task_state spawning. [ 943.154084] env[70013]: DEBUG nova.network.neutron [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.163245] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "refresh_cache-965480de-4fe2-48ce-94ad-0838d78239dd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.163492] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquired lock "refresh_cache-965480de-4fe2-48ce-94ad-0838d78239dd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.163681] env[70013]: DEBUG nova.network.neutron [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.165245] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.165575] env[70013]: DEBUG nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Instance network_info: |[{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 943.167516] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.167516] env[70013]: DEBUG nova.network.neutron [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing network info cache for port 3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.168014] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:3e:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e38f8cd-84a0-48be-9215-3dec9c9ced94', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.177370] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 943.182711] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.183451] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-671d9645-0528-4d08-a9fc-f6a5fa37fdde {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.202977] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d1ffdb-4437-4556-9567-cc2c60ae128d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.211169] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e306a95-9b4c-4836-a0df-943d6ebead49 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.216573] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.216573] env[70013]: value = "task-4231086" [ 943.216573] env[70013]: _type = "Task" [ 943.216573] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.246743] env[70013]: DEBUG nova.network.neutron [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.252242] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8bbb2e-33bf-4b14-aef1-99dc64963fe8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.260372] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231086, 'name': CreateVM_Task} progress is 15%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.267642] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1890f5b-7f45-47c5-9994-96f3c8880c74 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.285642] env[70013]: DEBUG nova.compute.provider_tree [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 943.332996] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5ec11ea9-d49a-481b-af9b-cd3ac77f8327 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.028s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.445998] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231085, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.462302] env[70013]: DEBUG nova.network.neutron [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updated VIF entry in instance network info cache for port 3e38f8cd-84a0-48be-9215-3dec9c9ced94. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.462683] env[70013]: DEBUG nova.network.neutron [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.526879] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.527317] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.527656] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.657706] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.695676] env[70013]: DEBUG nova.network.neutron [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.730850] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231086, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.760801] env[70013]: INFO nova.compute.manager [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] [instance: 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3] Took 1.05 seconds to deallocate network for instance. [ 943.770366] env[70013]: DEBUG nova.network.neutron [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.816853] env[70013]: ERROR nova.scheduler.client.report [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [req-db0408d2-d70a-4dc6-abe4-7ed42fc46758] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-db0408d2-d70a-4dc6-abe4-7ed42fc46758"}]} [ 943.817329] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.285s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.818157] env[70013]: ERROR nova.compute.manager [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Traceback (most recent call last): [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] yield [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] self.set_inventory_for_provider( [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-db0408d2-d70a-4dc6-abe4-7ed42fc46758"}]} [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] During handling of the above exception, another exception occurred: [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Traceback (most recent call last): [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] with self.rt.instance_claim(context, instance, node, allocs, [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] return f(*args, **kwargs) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] self._update(elevated, cn) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] self._update_to_placement(context, compute_node, startup) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] return attempt.get(self._wrap_exception) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] six.reraise(self.value[0], self.value[1], self.value[2]) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] raise value [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] self.reportclient.update_from_provider_tree( [ 943.818157] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 943.819115] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] with catch_all(pd.uuid): [ 943.819115] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 943.819115] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] self.gen.throw(typ, value, traceback) [ 943.819115] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 943.819115] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] raise exception.ResourceProviderSyncFailed() [ 943.819115] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 943.819115] env[70013]: ERROR nova.compute.manager [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] [ 943.819677] env[70013]: DEBUG nova.compute.utils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 943.821334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.992s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.823117] env[70013]: INFO nova.compute.claims [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.828782] env[70013]: DEBUG nova.compute.manager [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Build of instance dacc5f9c-3cd2-4992-b59a-260d2657567e was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 943.828782] env[70013]: DEBUG nova.compute.manager [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 943.828782] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-dacc5f9c-3cd2-4992-b59a-260d2657567e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.828782] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-dacc5f9c-3cd2-4992-b59a-260d2657567e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.828782] env[70013]: DEBUG nova.network.neutron [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.837315] env[70013]: DEBUG nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 943.945847] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231085, 'name': CloneVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.965708] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5a4987b-6244-407f-bb03-3d2c63b9c536 req-a1b2da3b-309c-4867-b2aa-d4b180fd46b1 service nova] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.071392] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 944.163389] env[70013]: DEBUG nova.compute.manager [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 944.164604] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53186650-3dae-4a72-b68d-a8cdf48a2851 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.232548] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231086, 'name': CreateVM_Task, 'duration_secs': 0.559932} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.233138] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 944.233503] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.233722] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.234117] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 944.234930] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32e15087-3fb6-4276-9d86-5767dfa3c48a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.241016] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 944.241016] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f9b982-19f7-1be8-c615-3e623f557346" [ 944.241016] env[70013]: _type = "Task" [ 944.241016] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.253703] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f9b982-19f7-1be8-c615-3e623f557346, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.273303] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Releasing lock "refresh_cache-965480de-4fe2-48ce-94ad-0838d78239dd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.273716] env[70013]: DEBUG nova.compute.manager [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 944.273910] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.274767] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ff0da4-f273-44c2-8f3f-677ec2c93c6a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.283017] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.283017] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ca30599-2866-46b9-9595-b288ce00bb56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.292969] env[70013]: DEBUG oslo_vmware.api [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 944.292969] env[70013]: value = "task-4231087" [ 944.292969] env[70013]: _type = "Task" [ 944.292969] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.306721] env[70013]: DEBUG oslo_vmware.api [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.349446] env[70013]: DEBUG nova.network.neutron [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 944.363369] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.448816] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231085, 'name': CloneVM_Task, 'duration_secs': 1.69055} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.449337] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Created linked-clone VM from snapshot [ 944.451843] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b464b0dc-759b-4412-bbd0-1567b502ba59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.459940] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Uploading image 645cd293-8a3c-429e-b6fa-19aca420f341 {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 944.468431] env[70013]: INFO nova.compute.manager [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Rebuilding instance [ 944.494298] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 944.494298] env[70013]: value = "vm-837152" [ 944.494298] env[70013]: _type = "VirtualMachine" [ 944.494298] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 944.496856] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-db2d6c84-4a78-4853-9ac5-55d7a226fd16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.506393] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lease: (returnval){ [ 944.506393] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c040cc-7b76-d94b-9f97-b75f02eb61d3" [ 944.506393] env[70013]: _type = "HttpNfcLease" [ 944.506393] env[70013]: } obtained for exporting VM: (result){ [ 944.506393] env[70013]: value = "vm-837152" [ 944.506393] env[70013]: _type = "VirtualMachine" [ 944.506393] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 944.507534] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the lease: (returnval){ [ 944.507534] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c040cc-7b76-d94b-9f97-b75f02eb61d3" [ 944.507534] env[70013]: _type = "HttpNfcLease" [ 944.507534] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 944.515200] env[70013]: DEBUG nova.network.neutron [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.526794] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 944.526794] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c040cc-7b76-d94b-9f97-b75f02eb61d3" [ 944.526794] env[70013]: _type = "HttpNfcLease" [ 944.526794] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 944.527396] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 944.527396] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c040cc-7b76-d94b-9f97-b75f02eb61d3" [ 944.527396] env[70013]: _type = "HttpNfcLease" [ 944.527396] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 944.528431] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450beee1-11aa-4b5f-88ac-b21555deae73 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.545015] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522cd9c6-c321-7401-6d21-4f0a88a6f22c/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 944.545306] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522cd9c6-c321-7401-6d21-4f0a88a6f22c/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 944.546992] env[70013]: DEBUG nova.compute.manager [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 944.547877] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be40d247-82fa-47ab-b154-d18b265aa286 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.652389] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c6ee5999-ce4c-46ed-afea-9b4b06c7b90c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.653928] env[70013]: DEBUG nova.network.neutron [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Updating instance_info_cache with network_info: [{"id": "6114ded0-69f6-4591-bc0f-defc91e44362", "address": "fa:16:3e:24:47:ce", "network": {"id": "3c2bbab7-386e-40f5-9bf8-fdb3a1d7d0e3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1633557551", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.116", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6114ded0-69", "ovs_interfaceid": "6114ded0-69f6-4591-bc0f-defc91e44362", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "address": "fa:16:3e:fd:3e:39", "network": {"id": "03d9009f-f8a8-4f1c-b205-36038ddbf3f1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-959921045", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd5389a-04", "ovs_interfaceid": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "address": "fa:16:3e:1f:7f:fb", "network": {"id": "3c2bbab7-386e-40f5-9bf8-fdb3a1d7d0e3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1633557551", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894fb3a3-f1", "ovs_interfaceid": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.755825] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f9b982-19f7-1be8-c615-3e623f557346, 'name': SearchDatastore_Task, 'duration_secs': 0.033329} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.756178] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.756421] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.756641] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.756791] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.756969] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.757342] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d6cfa17-3725-46fc-b63f-15ffe1064c58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.766691] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.766877] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 944.767618] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef7cbf98-e6ec-42ef-8975-a5cb455139b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.776203] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 944.776203] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ccb868-1bb1-4e55-d3fb-e8a17c5daa4e" [ 944.776203] env[70013]: _type = "Task" [ 944.776203] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.786698] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ccb868-1bb1-4e55-d3fb-e8a17c5daa4e, 'name': SearchDatastore_Task, 'duration_secs': 0.010529} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.788689] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a266d47-538e-47c4-894b-4aed3b886328 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.793257] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 944.793257] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5223c814-ed22-46eb-9e8f-c634df201fb0" [ 944.793257] env[70013]: _type = "Task" [ 944.793257] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.801505] env[70013]: INFO nova.scheduler.client.report [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Deleted allocations for instance 55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3 [ 944.819872] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5223c814-ed22-46eb-9e8f-c634df201fb0, 'name': SearchDatastore_Task, 'duration_secs': 0.010542} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.825376] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.825849] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6/1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.826498] env[70013]: DEBUG oslo_vmware.api [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231087, 'name': PowerOffVM_Task, 'duration_secs': 0.218558} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.826647] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e44e3110-f264-4d26-b63f-4d9a66b15424 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.829935] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.830160] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 944.834197] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2dd0b0d-e4e5-4a6c-bdd2-cab948fb4092 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.844988] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 944.844988] env[70013]: value = "task-4231090" [ 944.844988] env[70013]: _type = "Task" [ 944.844988] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.854940] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231090, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.863038] env[70013]: DEBUG nova.scheduler.client.report [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 944.865777] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 944.866055] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 944.866233] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Deleting the datastore file [datastore2] 965480de-4fe2-48ce-94ad-0838d78239dd {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.868875] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87c2c007-d6f6-46da-86dd-874a606a6256 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.875458] env[70013]: DEBUG oslo_vmware.api [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for the task: (returnval){ [ 944.875458] env[70013]: value = "task-4231091" [ 944.875458] env[70013]: _type = "Task" [ 944.875458] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.880520] env[70013]: DEBUG nova.scheduler.client.report [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 944.880777] env[70013]: DEBUG nova.compute.provider_tree [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 944.887081] env[70013]: DEBUG oslo_vmware.api [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.903308] env[70013]: DEBUG nova.scheduler.client.report [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 944.929288] env[70013]: DEBUG nova.scheduler.client.report [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 945.022498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-dacc5f9c-3cd2-4992-b59a-260d2657567e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.022746] env[70013]: DEBUG nova.compute.manager [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 945.022946] env[70013]: DEBUG nova.compute.manager [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 945.023162] env[70013]: DEBUG nova.network.neutron [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.045773] env[70013]: DEBUG nova.network.neutron [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.059282] env[70013]: DEBUG nova.compute.manager [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-changed-894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 945.059282] env[70013]: DEBUG nova.compute.manager [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Refreshing instance network info cache due to event network-changed-894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 945.059522] env[70013]: DEBUG oslo_concurrency.lockutils [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] Acquiring lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.158734] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Releasing lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.159526] env[70013]: DEBUG nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Instance network_info: |[{"id": "6114ded0-69f6-4591-bc0f-defc91e44362", "address": "fa:16:3e:24:47:ce", "network": {"id": "3c2bbab7-386e-40f5-9bf8-fdb3a1d7d0e3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1633557551", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.116", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6114ded0-69", "ovs_interfaceid": "6114ded0-69f6-4591-bc0f-defc91e44362", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "address": "fa:16:3e:fd:3e:39", "network": {"id": "03d9009f-f8a8-4f1c-b205-36038ddbf3f1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-959921045", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd5389a-04", "ovs_interfaceid": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "address": "fa:16:3e:1f:7f:fb", "network": {"id": "3c2bbab7-386e-40f5-9bf8-fdb3a1d7d0e3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1633557551", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894fb3a3-f1", "ovs_interfaceid": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 945.161362] env[70013]: DEBUG oslo_concurrency.lockutils [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] Acquired lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.161909] env[70013]: DEBUG nova.network.neutron [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Refreshing network info cache for port 894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 945.163590] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:47:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6114ded0-69f6-4591-bc0f-defc91e44362', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:3e:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604056d6-6dd6-47fa-9eaa-6863a3a7c488', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0dd5389a-041f-43b0-949a-78f2655dc2ee', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:7f:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.175358] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Creating folder: Project (4eef90bcd1794b788b1340d14c6df552). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 945.182419] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f12a9531-fef4-4db2-8518-6513f3b70416 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.185603] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ebe791-5a71-4c18-bf79-33b4b7984eb9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.202060] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Doing hard reboot of VM {{(pid=70013) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 945.205624] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0637a6bd-e4f9-4154-bb0c-e6318668b3b3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.209775] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Created folder: Project (4eef90bcd1794b788b1340d14c6df552) in parent group-v836999. [ 945.211377] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Creating folder: Instances. Parent ref: group-v837154. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 945.211377] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f54f13d-c55d-4fba-8acc-db9176fb448d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.217203] env[70013]: DEBUG oslo_vmware.api [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 945.217203] env[70013]: value = "task-4231093" [ 945.217203] env[70013]: _type = "Task" [ 945.217203] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.228711] env[70013]: DEBUG oslo_vmware.api [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231093, 'name': ResetVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.233923] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Created folder: Instances in parent group-v837154. [ 945.234311] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 945.234949] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.235215] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77d9033e-0d42-4fa4-9a26-356ad97c9f3c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.274857] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.274857] env[70013]: value = "task-4231095" [ 945.274857] env[70013]: _type = "Task" [ 945.274857] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.290071] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231095, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.325309] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02ca5b37-75c5-4712-bbd8-c6f559c82435 tempest-ServerMetadataTestJSON-1933974841 tempest-ServerMetadataTestJSON-1933974841-project-member] Lock "55ea6a53-fe4b-4b21-ab6f-4eb009ad69c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.212s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 945.367580] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231090, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.393503] env[70013]: DEBUG oslo_vmware.api [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Task: {'id': task-4231091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126428} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.395472] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.395472] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.395472] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.395688] env[70013]: INFO nova.compute.manager [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Took 1.12 seconds to destroy the instance on the hypervisor. [ 945.395954] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 945.400044] env[70013]: DEBUG nova.compute.manager [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 945.400044] env[70013]: DEBUG nova.network.neutron [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.431024] env[70013]: DEBUG nova.network.neutron [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.537241] env[70013]: DEBUG nova.network.neutron [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Updated VIF entry in instance network info cache for port 894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.538700] env[70013]: DEBUG nova.network.neutron [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Updating instance_info_cache with network_info: [{"id": "6114ded0-69f6-4591-bc0f-defc91e44362", "address": "fa:16:3e:24:47:ce", "network": {"id": "3c2bbab7-386e-40f5-9bf8-fdb3a1d7d0e3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1633557551", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.116", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6114ded0-69", "ovs_interfaceid": "6114ded0-69f6-4591-bc0f-defc91e44362", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "address": "fa:16:3e:fd:3e:39", "network": {"id": "03d9009f-f8a8-4f1c-b205-36038ddbf3f1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-959921045", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd5389a-04", "ovs_interfaceid": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "address": "fa:16:3e:1f:7f:fb", "network": {"id": "3c2bbab7-386e-40f5-9bf8-fdb3a1d7d0e3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1633557551", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894fb3a3-f1", "ovs_interfaceid": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.543652] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120cd37b-b4b3-4b99-86e7-a57f06b592ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.552307] env[70013]: DEBUG nova.network.neutron [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.563028] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6871b20f-dfa7-4958-9b7f-747810a9156c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.605264] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5dacd31-19e0-4ead-9206-98d7d0a957d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.614444] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5b9dd1-74d8-47a5-87aa-ffe03d185044 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.631969] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.632601] env[70013]: DEBUG nova.compute.provider_tree [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.634314] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7c23762-e358-498d-a3d6-df0d96abe289 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.644012] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 945.644012] env[70013]: value = "task-4231096" [ 945.644012] env[70013]: _type = "Task" [ 945.644012] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.655064] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.729668] env[70013]: DEBUG oslo_vmware.api [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231093, 'name': ResetVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.788749] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231095, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.832694] env[70013]: DEBUG nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 945.862182] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231090, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562531} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.862510] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6/1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.862763] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.863085] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93008dd2-2182-4777-b253-d86abd997acc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.871552] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 945.871552] env[70013]: value = "task-4231097" [ 945.871552] env[70013]: _type = "Task" [ 945.871552] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.885654] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231097, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.932163] env[70013]: DEBUG nova.network.neutron [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.043092] env[70013]: DEBUG oslo_concurrency.lockutils [req-18f515e2-ca0d-44d3-bf4e-0d34eb1277b8 req-22fdfca2-7ab6-493d-8ff6-93031a9ff96b service nova] Releasing lock "refresh_cache-ab091ef4-9828-438f-8e95-4370bfbd34de" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.056801] env[70013]: INFO nova.compute.manager [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: dacc5f9c-3cd2-4992-b59a-260d2657567e] Took 1.03 seconds to deallocate network for instance. [ 946.141027] env[70013]: DEBUG nova.scheduler.client.report [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.159856] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231096, 'name': PowerOffVM_Task, 'duration_secs': 0.249356} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.160217] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.160529] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.161410] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948473b2-577d-4080-9be4-e30b4ecb8860 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.172367] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.172752] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-baa9c741-8047-4f78-825b-24fe5dbb7ba6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.228863] env[70013]: DEBUG oslo_vmware.api [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231093, 'name': ResetVM_Task, 'duration_secs': 0.701871} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.229158] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Did hard reboot of VM {{(pid=70013) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 946.229358] env[70013]: DEBUG nova.compute.manager [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.230186] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3b596e-fcd7-4bd2-a4a7-b7374925de17 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.247252] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.247252] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.247829] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleting the datastore file [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.247829] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fd24890-8b42-428e-8bc7-caeb1227f47e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.256240] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 946.256240] env[70013]: value = "task-4231099" [ 946.256240] env[70013]: _type = "Task" [ 946.256240] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.266784] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231099, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.287598] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231095, 'name': CreateVM_Task, 'duration_secs': 0.600939} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.288586] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.289543] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.289710] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.290049] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 946.290687] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33c5944-ffe7-4f81-9f30-d940c04dbf4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.297195] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 946.297195] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200fd10-6f1d-5faf-8b54-da3c3d32fde4" [ 946.297195] env[70013]: _type = "Task" [ 946.297195] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.308167] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200fd10-6f1d-5faf-8b54-da3c3d32fde4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.358591] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.382066] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231097, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081474} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.382463] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.383439] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5aec70c-c370-4b48-b445-d7cb408a2e08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.409673] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6/1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.410181] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84a990f4-436b-4b31-bfa6-97e5a06953f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.433808] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 946.433808] env[70013]: value = "task-4231100" [ 946.433808] env[70013]: _type = "Task" [ 946.433808] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.444076] env[70013]: INFO nova.compute.manager [-] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Took 1.05 seconds to deallocate network for instance. [ 946.444536] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.649256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.828s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.649859] env[70013]: DEBUG nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 946.652757] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.041s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.656021] env[70013]: DEBUG nova.objects.instance [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lazy-loading 'resources' on Instance uuid 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.747611] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2e003402-69f1-4e92-b1f4-db6093123e3b tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.463s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.769508] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231099, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.374383} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.769508] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.769508] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.769508] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.814521] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5200fd10-6f1d-5faf-8b54-da3c3d32fde4, 'name': SearchDatastore_Task, 'duration_secs': 0.031529} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.814852] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.815267] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.815420] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.815598] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.815822] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.816446] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e755a9e-3145-4a1d-aa0c-a3c0d8d1e917 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.827399] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.827650] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.828605] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed93f018-5868-481b-92b5-15d505c7d29a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.835785] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 946.835785] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52aedd30-d779-11d9-7215-9ace3409e214" [ 946.835785] env[70013]: _type = "Task" [ 946.835785] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.845172] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52aedd30-d779-11d9-7215-9ace3409e214, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.945408] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231100, 'name': ReconfigVM_Task, 'duration_secs': 0.416347} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.945731] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6/1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.946638] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d681823-a9f5-4c13-bbae-e86e7f56b8dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.953985] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.961485] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 946.961485] env[70013]: value = "task-4231101" [ 946.961485] env[70013]: _type = "Task" [ 946.961485] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.971457] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231101, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.159380] env[70013]: DEBUG nova.compute.utils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 947.163967] env[70013]: DEBUG nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 947.163967] env[70013]: DEBUG nova.network.neutron [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.231111] env[70013]: INFO nova.scheduler.client.report [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted allocations for instance dacc5f9c-3cd2-4992-b59a-260d2657567e [ 947.255475] env[70013]: DEBUG nova.policy [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7edeb687382d47e59b9f07bd1397e563', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ca0b60f856f4cd7a243629fcee1cd42', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 947.354548] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52aedd30-d779-11d9-7215-9ace3409e214, 'name': SearchDatastore_Task, 'duration_secs': 0.015383} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.360203] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4991694e-d8e0-4a6d-a82f-e85993d3d960 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.368301] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 947.368301] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d17e84-9cec-8865-7fac-26af624405f4" [ 947.368301] env[70013]: _type = "Task" [ 947.368301] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.382989] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d17e84-9cec-8865-7fac-26af624405f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.474525] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231101, 'name': Rename_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.666560] env[70013]: DEBUG nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 947.725251] env[70013]: DEBUG nova.network.neutron [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Successfully created port: 7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 947.729104] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b70da0e-f20f-43ce-89e2-51115f2b185a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.739775] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d7b25a-2ac5-459a-8043-e035a46314ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.746380] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6ec70484-5dd2-4287-bfbc-7ed99287040c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "dacc5f9c-3cd2-4992-b59a-260d2657567e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.399s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.785954] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001fcff6-78b8-4446-a93c-83248ce98a3d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.798605] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650b10b1-7886-4d8e-b568-cafa4c7bf17a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.817484] env[70013]: DEBUG nova.compute.provider_tree [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.829746] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 947.830076] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.830284] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 947.830744] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.830954] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 947.831115] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 947.831393] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 947.831681] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 947.831889] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 947.832081] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 947.832328] env[70013]: DEBUG nova.virt.hardware [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 947.834273] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a93bf1-dc7a-46ff-9164-ae12bd54e5a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.845437] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e43f70-28b9-4ed8-a3e5-69ba7727e942 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.862773] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:c1:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a55f45a-d631-4ebc-b73b-8a30bd0a32a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6f11372-627a-4dc6-8fe9-e20fcc0a36ed', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 947.871059] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 947.871481] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 947.875412] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94d75397-cbff-4f05-9adb-a23456e8a212 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.899215] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d17e84-9cec-8865-7fac-26af624405f4, 'name': SearchDatastore_Task, 'duration_secs': 0.014735} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.901207] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.901567] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] ab091ef4-9828-438f-8e95-4370bfbd34de/ab091ef4-9828-438f-8e95-4370bfbd34de.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.901790] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.901790] env[70013]: value = "task-4231102" [ 947.901790] env[70013]: _type = "Task" [ 947.901790] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.901972] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ade7e8b-24b8-44ee-980e-d01e8b6c30fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.914046] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231102, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.915909] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 947.915909] env[70013]: value = "task-4231103" [ 947.915909] env[70013]: _type = "Task" [ 947.915909] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.925672] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.975223] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231101, 'name': Rename_Task, 'duration_secs': 0.999907} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.975467] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.975903] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5d85705-b805-4357-bf25-865365ea1276 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.984299] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 947.984299] env[70013]: value = "task-4231104" [ 947.984299] env[70013]: _type = "Task" [ 947.984299] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.996080] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231104, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.285508] env[70013]: DEBUG nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 948.325315] env[70013]: DEBUG nova.scheduler.client.report [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.417034] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231102, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.426278] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231103, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.496082] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231104, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.676650] env[70013]: DEBUG nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 948.705192] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 948.705551] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.705825] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 948.706120] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.706361] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 948.706579] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 948.706849] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 948.707091] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 948.707318] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 948.707558] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 948.707802] env[70013]: DEBUG nova.virt.hardware [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 948.709142] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43dc6ba-8639-4a60-80da-9049de5b4ea5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.720371] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f745cb2-f72f-409c-a5a3-b46c1bc3cb4b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.817709] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.830050] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.177s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.833218] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.492s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.834055] env[70013]: DEBUG nova.objects.instance [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lazy-loading 'resources' on Instance uuid d84d8426-b3fd-4f0d-8410-0ee24fedb35d {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.855625] env[70013]: INFO nova.scheduler.client.report [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Deleted allocations for instance 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5 [ 948.918583] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231102, 'name': CreateVM_Task, 'duration_secs': 0.755403} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.922363] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.923720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.923720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 948.923720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 948.924463] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-698c8757-edc1-4499-b593-897321d4b732 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.930612] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695651} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.931329] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] ab091ef4-9828-438f-8e95-4370bfbd34de/ab091ef4-9828-438f-8e95-4370bfbd34de.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.931480] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.932040] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-666d7abf-9483-4699-a35d-d9422d7ce2e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.935717] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 948.935717] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211f06c-4bba-f878-849d-f1dec56ae6dd" [ 948.935717] env[70013]: _type = "Task" [ 948.935717] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.943567] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 948.943567] env[70013]: value = "task-4231105" [ 948.943567] env[70013]: _type = "Task" [ 948.943567] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.952182] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211f06c-4bba-f878-849d-f1dec56ae6dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.960988] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231105, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.996031] env[70013]: DEBUG oslo_vmware.api [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231104, 'name': PowerOnVM_Task, 'duration_secs': 0.764741} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.996353] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.996625] env[70013]: INFO nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Took 9.82 seconds to spawn the instance on the hypervisor. [ 948.996938] env[70013]: DEBUG nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 948.997895] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85f951e-6087-46af-b327-61abeafc166c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.318158] env[70013]: DEBUG nova.compute.manager [req-7c93179d-05c0-4c68-90bc-ca66328abb3f req-7168af62-8fee-4d82-ad4e-357c88548c1e service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received event network-vif-plugged-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 949.318419] env[70013]: DEBUG oslo_concurrency.lockutils [req-7c93179d-05c0-4c68-90bc-ca66328abb3f req-7168af62-8fee-4d82-ad4e-357c88548c1e service nova] Acquiring lock "39557c50-866a-4510-b840-b1a6a3e3890e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.318732] env[70013]: DEBUG oslo_concurrency.lockutils [req-7c93179d-05c0-4c68-90bc-ca66328abb3f req-7168af62-8fee-4d82-ad4e-357c88548c1e service nova] Lock "39557c50-866a-4510-b840-b1a6a3e3890e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.318860] env[70013]: DEBUG oslo_concurrency.lockutils [req-7c93179d-05c0-4c68-90bc-ca66328abb3f req-7168af62-8fee-4d82-ad4e-357c88548c1e service nova] Lock "39557c50-866a-4510-b840-b1a6a3e3890e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.319267] env[70013]: DEBUG nova.compute.manager [req-7c93179d-05c0-4c68-90bc-ca66328abb3f req-7168af62-8fee-4d82-ad4e-357c88548c1e service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] No waiting events found dispatching network-vif-plugged-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 949.319486] env[70013]: WARNING nova.compute.manager [req-7c93179d-05c0-4c68-90bc-ca66328abb3f req-7168af62-8fee-4d82-ad4e-357c88548c1e service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received unexpected event network-vif-plugged-7beab378-23cf-4f47-8004-b917e0a1e482 for instance with vm_state building and task_state spawning. [ 949.367245] env[70013]: DEBUG oslo_concurrency.lockutils [None req-17e72379-b537-4ea5-b4fe-df1053edec3d tempest-VolumesAdminNegativeTest-932475284 tempest-VolumesAdminNegativeTest-932475284-project-member] Lock "8ec4b86d-e83b-4b28-ac4b-d990177b0ba5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.635s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.418613] env[70013]: DEBUG nova.network.neutron [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Successfully updated port: 7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.428309] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "2070989d-6c90-4eb1-8508-7587ed9659d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.428613] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.453736] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211f06c-4bba-f878-849d-f1dec56ae6dd, 'name': SearchDatastore_Task, 'duration_secs': 0.022072} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.453957] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.454662] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.454662] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.454662] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.454804] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.455933] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a797fe8-deb2-44cd-a8ff-b7ef550fbb00 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.460817] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231105, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075678} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.463757] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.464840] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57949c42-aa07-4966-8445-956eb7b91d26 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.494635] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] ab091ef4-9828-438f-8e95-4370bfbd34de/ab091ef4-9828-438f-8e95-4370bfbd34de.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.500274] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-960c70d2-fbba-4ec1-8852-858dbe2e282c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.516343] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 949.516680] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 949.523101] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d3ce69e-16d7-4266-a225-79c68cfe7b90 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.529223] env[70013]: INFO nova.compute.manager [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Took 45.45 seconds to build instance. [ 949.531697] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 949.531697] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5254080d-30d1-bde3-8f0b-f35f28dc03a9" [ 949.531697] env[70013]: _type = "Task" [ 949.531697] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.539137] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 949.539137] env[70013]: value = "task-4231106" [ 949.539137] env[70013]: _type = "Task" [ 949.539137] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.550180] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5254080d-30d1-bde3-8f0b-f35f28dc03a9, 'name': SearchDatastore_Task, 'duration_secs': 0.015081} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.553145] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90fa06b4-dc5b-4f98-a4b7-d3bdc5e8a659 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.560426] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231106, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.569019] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 949.569019] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5272fd6f-b623-0f88-39ad-cf87c83b252e" [ 949.569019] env[70013]: _type = "Task" [ 949.569019] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.576893] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5272fd6f-b623-0f88-39ad-cf87c83b252e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.854098] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b3b444-bb6b-491d-8100-6e69c58b068a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.865036] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcba56da-fddd-4463-9cd2-9a93fc5f1d33 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.899578] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d98c1c-6e6d-48c4-bc5d-bd6e8bf9e3ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.908985] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea471ef-d409-4e9e-af39-1343fc2bec66 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.925324] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.925446] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquired lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.925633] env[70013]: DEBUG nova.network.neutron [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 949.927056] env[70013]: DEBUG nova.compute.provider_tree [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.033292] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ae2b6fe0-0ac6-4815-af31-2750fd7eb532 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.986s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.054756] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231106, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.079129] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5272fd6f-b623-0f88-39ad-cf87c83b252e, 'name': SearchDatastore_Task, 'duration_secs': 0.020819} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.079432] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.079708] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 950.080010] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-19811d9a-ace7-4f8e-ad80-a71a0de492e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.089238] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 950.089238] env[70013]: value = "task-4231107" [ 950.089238] env[70013]: _type = "Task" [ 950.089238] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.100387] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231107, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.432754] env[70013]: DEBUG nova.scheduler.client.report [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 950.492382] env[70013]: DEBUG nova.network.neutron [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 950.539202] env[70013]: DEBUG nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 950.553778] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231106, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.602919] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231107, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.937922] env[70013]: DEBUG nova.network.neutron [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.944113] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.110s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.949255] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.551s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.949562] env[70013]: DEBUG nova.objects.instance [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lazy-loading 'resources' on Instance uuid 2aefaa91-c439-486a-8b19-c6f45f52583f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.978622] env[70013]: INFO nova.scheduler.client.report [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Deleted allocations for instance d84d8426-b3fd-4f0d-8410-0ee24fedb35d [ 951.056204] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231106, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.059751] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.101146] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231107, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640778} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.101435] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 951.101675] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.101937] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25bab6df-9043-4a8e-a3e4-b0ae4c4d9669 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.110664] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 951.110664] env[70013]: value = "task-4231108" [ 951.110664] env[70013]: _type = "Task" [ 951.110664] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.121549] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231108, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.346836] env[70013]: DEBUG nova.compute.manager [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 951.347071] env[70013]: DEBUG nova.compute.manager [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing instance network info cache due to event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 951.347270] env[70013]: DEBUG oslo_concurrency.lockutils [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] Acquiring lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.443635] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Releasing lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.444066] env[70013]: DEBUG nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Instance network_info: |[{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 951.444400] env[70013]: DEBUG oslo_concurrency.lockutils [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] Acquired lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.444580] env[70013]: DEBUG nova.network.neutron [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.446295] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:6d:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7beab378-23cf-4f47-8004-b917e0a1e482', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.456411] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Creating folder: Project (7ca0b60f856f4cd7a243629fcee1cd42). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 951.463151] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfde42f0-6a0e-4e6b-8b81-3d8de13faab0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.477337] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Created folder: Project (7ca0b60f856f4cd7a243629fcee1cd42) in parent group-v836999. [ 951.478267] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Creating folder: Instances. Parent ref: group-v837158. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 951.478267] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eaa19cf3-dc41-4395-be4e-fa3bb56e1bd2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.487899] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3e04809d-1f13-4f6d-82ba-116451981839 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "d84d8426-b3fd-4f0d-8410-0ee24fedb35d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.752s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.499036] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Created folder: Instances in parent group-v837158. [ 951.499036] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 951.499713] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.499834] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcb775a9-238b-4a44-9681-9c4c2e89722a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.533044] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.533044] env[70013]: value = "task-4231111" [ 951.533044] env[70013]: _type = "Task" [ 951.533044] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.547150] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231111, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.560382] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231106, 'name': ReconfigVM_Task, 'duration_secs': 1.621189} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.560649] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Reconfigured VM instance instance-0000003f to attach disk [datastore2] ab091ef4-9828-438f-8e95-4370bfbd34de/ab091ef4-9828-438f-8e95-4370bfbd34de.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.561356] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6074c7dd-2e1c-4a10-a72c-a8314482ac08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.569867] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 951.569867] env[70013]: value = "task-4231112" [ 951.569867] env[70013]: _type = "Task" [ 951.569867] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.584333] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231112, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.622531] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231108, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079058} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.625497] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.626738] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a913e7-a140-4b9e-8b35-62f7f284b862 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.655997] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.661640] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6ad9e96-f884-4824-8f27-9b1f42838316 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.685050] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 951.685050] env[70013]: value = "task-4231113" [ 951.685050] env[70013]: _type = "Task" [ 951.685050] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.699106] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231113, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.904683] env[70013]: DEBUG nova.network.neutron [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updated VIF entry in instance network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 951.904683] env[70013]: DEBUG nova.network.neutron [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.047704] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231111, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.051509] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf00a0f-05dd-43f7-9cb1-a52a6269528e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.061428] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd4372d-56e6-439b-a3d8-e4ad8ed33b1b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.097240] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be814789-40fb-457c-bdac-960d930e9005 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.108729] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce29c18-e55f-48c1-892c-62c8db157136 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.113237] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231112, 'name': Rename_Task, 'duration_secs': 0.256766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.113237] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.113442] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-931f4145-bea6-425c-a5ef-5541c069e54d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.126426] env[70013]: DEBUG nova.compute.provider_tree [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 952.129331] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 952.129331] env[70013]: value = "task-4231114" [ 952.129331] env[70013]: _type = "Task" [ 952.129331] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.138674] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231114, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.197914] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.409172] env[70013]: DEBUG oslo_concurrency.lockutils [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] Releasing lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.409172] env[70013]: DEBUG nova.compute.manager [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-changed-3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 952.409172] env[70013]: DEBUG nova.compute.manager [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing instance network info cache due to event network-changed-3e38f8cd-84a0-48be-9215-3dec9c9ced94. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 952.409321] env[70013]: DEBUG oslo_concurrency.lockutils [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.409422] env[70013]: DEBUG oslo_concurrency.lockutils [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.409599] env[70013]: DEBUG nova.network.neutron [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing network info cache for port 3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.561783] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231111, 'name': CreateVM_Task, 'duration_secs': 0.520138} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.562182] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.563496] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.563980] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.564552] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 952.565044] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4da231f-487f-4915-a5ae-2d38f54fd0b1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.573529] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 952.573529] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52253182-a4f6-7a2e-98e3-9e23cec55c3a" [ 952.573529] env[70013]: _type = "Task" [ 952.573529] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.590763] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52253182-a4f6-7a2e-98e3-9e23cec55c3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.645526] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231114, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.665571] env[70013]: ERROR nova.scheduler.client.report [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [req-6cf2dd0b-e8f3-4eb5-b1b8-77cbbb05a812] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6cf2dd0b-e8f3-4eb5-b1b8-77cbbb05a812"}]} [ 952.666221] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.666969] env[70013]: ERROR nova.compute.manager [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Traceback (most recent call last): [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] yield [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self.set_inventory_for_provider( [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6cf2dd0b-e8f3-4eb5-b1b8-77cbbb05a812"}]} [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] During handling of the above exception, another exception occurred: [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Traceback (most recent call last): [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self._delete_instance(context, instance, bdms) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self._complete_deletion(context, instance) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self._update_resource_tracker(context, instance) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self.rt.update_usage(context, instance, instance.node) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] return f(*args, **kwargs) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self._update(context.elevated(), self.compute_nodes[nodename]) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self._update_to_placement(context, compute_node, startup) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] return attempt.get(self._wrap_exception) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] six.reraise(self.value[0], self.value[1], self.value[2]) [ 952.666969] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] raise value [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self.reportclient.update_from_provider_tree( [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] with catch_all(pd.uuid): [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] self.gen.throw(typ, value, traceback) [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] raise exception.ResourceProviderSyncFailed() [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 952.668086] env[70013]: ERROR nova.compute.manager [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] [ 952.670957] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.430s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.677825] env[70013]: INFO nova.compute.claims [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.697632] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231113, 'name': ReconfigVM_Task, 'duration_secs': 0.728144} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.697960] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df/626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.698708] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d24453af-bc05-4289-ad7c-4e311e512636 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.706539] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 952.706539] env[70013]: value = "task-4231115" [ 952.706539] env[70013]: _type = "Task" [ 952.706539] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.716749] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231115, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.095891] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52253182-a4f6-7a2e-98e3-9e23cec55c3a, 'name': SearchDatastore_Task, 'duration_secs': 0.025449} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.095891] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.095891] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.096259] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.096296] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.097106] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.097106] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-139156a7-168d-4c5f-812e-02c4090f0751 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.119684] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.119889] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.120771] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9366e37-4382-4312-825f-1458c35e7512 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.128266] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 953.128266] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca9c50-174a-c673-e119-664b6d090454" [ 953.128266] env[70013]: _type = "Task" [ 953.128266] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.137761] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca9c50-174a-c673-e119-664b6d090454, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.145918] env[70013]: DEBUG oslo_vmware.api [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231114, 'name': PowerOnVM_Task, 'duration_secs': 0.576359} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.146232] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.146404] env[70013]: INFO nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Took 16.67 seconds to spawn the instance on the hypervisor. [ 953.146587] env[70013]: DEBUG nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 953.147361] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b9fc87-7dec-415c-b613-bc295ed67775 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.188189] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.285s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.219953] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231115, 'name': Rename_Task, 'duration_secs': 0.356862} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.220555] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.220918] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3132f646-40f4-49b9-8232-7805b7fda124 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.228924] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 953.228924] env[70013]: value = "task-4231116" [ 953.228924] env[70013]: _type = "Task" [ 953.228924] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.241829] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.273050] env[70013]: DEBUG nova.network.neutron [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updated VIF entry in instance network info cache for port 3e38f8cd-84a0-48be-9215-3dec9c9ced94. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.273050] env[70013]: DEBUG nova.network.neutron [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.641025] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca9c50-174a-c673-e119-664b6d090454, 'name': SearchDatastore_Task, 'duration_secs': 0.029399} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.641025] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01370a53-8211-48df-a0b5-cbcb639e468a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.648865] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 953.648865] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529098c3-c896-d8fd-86a4-c479784ab7b7" [ 953.648865] env[70013]: _type = "Task" [ 953.648865] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.658540] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529098c3-c896-d8fd-86a4-c479784ab7b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.671501] env[70013]: INFO nova.compute.manager [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Took 52.27 seconds to build instance. [ 953.716989] env[70013]: DEBUG nova.scheduler.client.report [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 953.731664] env[70013]: DEBUG nova.scheduler.client.report [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 953.731999] env[70013]: DEBUG nova.compute.provider_tree [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 953.757388] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231116, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.758990] env[70013]: DEBUG nova.scheduler.client.report [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 953.779830] env[70013]: DEBUG nova.scheduler.client.report [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 953.787153] env[70013]: DEBUG oslo_concurrency.lockutils [req-55a70884-3046-4f65-b24f-a4c62fd15b37 req-bfb14793-b89b-493e-8e41-7d86d1af0a07 service nova] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.161614] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529098c3-c896-d8fd-86a4-c479784ab7b7, 'name': SearchDatastore_Task, 'duration_secs': 0.015668} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.162839] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.162839] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/39557c50-866a-4510-b840-b1a6a3e3890e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.162839] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0f56ac7-f94b-42a0-924c-9a5bb78f3e9a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.174178] env[70013]: DEBUG oslo_concurrency.lockutils [None req-bc08f249-adda-40b2-a374-462dbecadbaa tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.553s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.181883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "69f67e8d-2da0-474c-92db-8388c98b61f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.182344] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "69f67e8d-2da0-474c-92db-8388c98b61f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.184450] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 954.184450] env[70013]: value = "task-4231117" [ 954.184450] env[70013]: _type = "Task" [ 954.184450] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.205107] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.218870] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "5a535178-36b5-4015-9cd4-1a3b9b7517c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.219205] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "5a535178-36b5-4015-9cd4-1a3b9b7517c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.249485] env[70013]: DEBUG oslo_vmware.api [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231116, 'name': PowerOnVM_Task, 'duration_secs': 1.017801} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.252333] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.252988] env[70013]: DEBUG nova.compute.manager [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 954.254484] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a769de75-3757-4754-bdda-00874ef8e24e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.295317] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d41d85-7577-4706-818d-f7f03fb91e1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.307514] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100b6efe-cb0e-407d-9c07-8a317f3fa715 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.343496] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6434c01-edc3-4eb7-b56e-c217d7c48360 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.358755] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f9f16c-b5d3-42a9-9a9d-a5ed40018824 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.375921] env[70013]: DEBUG nova.compute.provider_tree [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.686572] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 954.700107] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231117, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.709161] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.775457] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.853858] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "ab091ef4-9828-438f-8e95-4370bfbd34de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.854375] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.854485] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.854678] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.854850] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.857514] env[70013]: INFO nova.compute.manager [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Terminating instance [ 954.901841] env[70013]: ERROR nova.scheduler.client.report [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [req-1be9f22e-cefa-45df-af8d-8e4105cd7863] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1be9f22e-cefa-45df-af8d-8e4105cd7863"}]} [ 954.902197] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.902876] env[70013]: ERROR nova.compute.manager [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Traceback (most recent call last): [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] yield [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] self.set_inventory_for_provider( [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1be9f22e-cefa-45df-af8d-8e4105cd7863"}]} [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] During handling of the above exception, another exception occurred: [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Traceback (most recent call last): [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] with self.rt.instance_claim(context, instance, node, allocs, [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] return f(*args, **kwargs) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] self._update(elevated, cn) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] self._update_to_placement(context, compute_node, startup) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] return attempt.get(self._wrap_exception) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] six.reraise(self.value[0], self.value[1], self.value[2]) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] raise value [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] self.reportclient.update_from_provider_tree( [ 954.902876] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 954.903668] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] with catch_all(pd.uuid): [ 954.903668] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 954.903668] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] self.gen.throw(typ, value, traceback) [ 954.903668] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 954.903668] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] raise exception.ResourceProviderSyncFailed() [ 954.903668] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 954.903668] env[70013]: ERROR nova.compute.manager [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] [ 954.903668] env[70013]: DEBUG nova.compute.utils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 954.904754] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.121s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.904971] env[70013]: DEBUG nova.objects.instance [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lazy-loading 'resources' on Instance uuid 2c0e4326-c33a-42bb-b793-a100157b1c03 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.906579] env[70013]: DEBUG nova.compute.manager [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Build of instance 8d28f827-c773-4f49-b89d-7d3dd8df4e70 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 954.906985] env[70013]: DEBUG nova.compute.manager [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 954.908394] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "refresh_cache-8d28f827-c773-4f49-b89d-7d3dd8df4e70" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.908558] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquired lock "refresh_cache-8d28f827-c773-4f49-b89d-7d3dd8df4e70" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.908745] env[70013]: DEBUG nova.network.neutron [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.964136] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522cd9c6-c321-7401-6d21-4f0a88a6f22c/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 954.965119] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0173dc94-e677-4317-8e43-cb6e33006b3b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.971968] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522cd9c6-c321-7401-6d21-4f0a88a6f22c/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 954.972603] env[70013]: ERROR oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522cd9c6-c321-7401-6d21-4f0a88a6f22c/disk-0.vmdk due to incomplete transfer. [ 954.972869] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cd4a869d-d8b9-4546-958e-70ce867c4934 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.981436] env[70013]: DEBUG oslo_vmware.rw_handles [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522cd9c6-c321-7401-6d21-4f0a88a6f22c/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 954.981660] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Uploaded image 645cd293-8a3c-429e-b6fa-19aca420f341 to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 954.984163] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 954.984442] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6b0361b7-f1bd-4c6e-9d06-9135c5e10ffd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.991941] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 954.991941] env[70013]: value = "task-4231118" [ 954.991941] env[70013]: _type = "Task" [ 954.991941] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.000602] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231118, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.204249] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231117, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.831144} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.205595] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/39557c50-866a-4510-b840-b1a6a3e3890e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.205783] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.208156] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c69e4dc4-0974-49e5-aead-20a9a003c93d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.216718] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.221525] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 955.221525] env[70013]: value = "task-4231119" [ 955.221525] env[70013]: _type = "Task" [ 955.221525] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.237079] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231119, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.362804] env[70013]: DEBUG nova.compute.manager [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 955.362804] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.363358] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f029c1d7-af36-424e-be31-96090806e123 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.375543] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.375932] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dac2bb15-5350-4309-ba67-ef99970940ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.385403] env[70013]: DEBUG oslo_vmware.api [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 955.385403] env[70013]: value = "task-4231120" [ 955.385403] env[70013]: _type = "Task" [ 955.385403] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.398809] env[70013]: DEBUG oslo_vmware.api [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.431665] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 955.436116] env[70013]: DEBUG nova.network.neutron [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 955.448148] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 955.448338] env[70013]: DEBUG nova.compute.provider_tree [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 955.468211] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 955.497958] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 955.507201] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231118, 'name': Destroy_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.570158] env[70013]: DEBUG nova.network.neutron [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.713343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.713343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.713343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.713343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.713343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.715091] env[70013]: INFO nova.compute.manager [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Terminating instance [ 955.737465] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231119, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078536} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.737750] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.738821] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0dc9ae1-cd87-4c6c-85cf-15f2a282beb9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.766437] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/39557c50-866a-4510-b840-b1a6a3e3890e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.769480] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ce48ed6-cad6-41c0-bcef-e8bfe0621a22 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.790689] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 955.790689] env[70013]: value = "task-4231121" [ 955.790689] env[70013]: _type = "Task" [ 955.790689] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.802846] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231121, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.898431] env[70013]: DEBUG oslo_vmware.api [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231120, 'name': PowerOffVM_Task, 'duration_secs': 0.276016} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.898662] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.898831] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.900120] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab43bf23-fbcb-4323-aff0-b1b2e58a1694 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.004052] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231118, 'name': Destroy_Task, 'duration_secs': 0.989626} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.005976] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Destroyed the VM [ 956.006258] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 956.007699] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-677ef51b-78b0-442b-9c91-6fb61c198b22 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.015974] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 956.015974] env[70013]: value = "task-4231123" [ 956.015974] env[70013]: _type = "Task" [ 956.015974] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.030540] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231123, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.036743] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c532e3c0-b4dd-4e7d-af4c-fc4bd9587122 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.045910] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e659b3-919a-47b2-a3e1-ae5c2a2a9b3b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.078581] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Releasing lock "refresh_cache-8d28f827-c773-4f49-b89d-7d3dd8df4e70" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 956.078788] env[70013]: DEBUG nova.compute.manager [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 956.079027] env[70013]: DEBUG nova.compute.manager [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 956.079266] env[70013]: DEBUG nova.network.neutron [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.081989] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d9f605-fc44-49f7-b0bb-3359b10084d4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.091429] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5144f8e0-7a3f-4594-a9fa-13523eed9bec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.106021] env[70013]: DEBUG nova.compute.provider_tree [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 956.108501] env[70013]: DEBUG nova.network.neutron [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.111281] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.111539] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.111704] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Deleting the datastore file [datastore2] ab091ef4-9828-438f-8e95-4370bfbd34de {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.112230] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e31e21d-2c9e-436e-af9e-22c119e1dccc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.119568] env[70013]: DEBUG oslo_vmware.api [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 956.119568] env[70013]: value = "task-4231124" [ 956.119568] env[70013]: _type = "Task" [ 956.119568] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.129165] env[70013]: DEBUG oslo_vmware.api [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.220067] env[70013]: DEBUG nova.compute.manager [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 956.220321] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.221320] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0268c8e2-0865-4afc-a13e-cf25f0f5bd86 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.231648] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.231948] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a3c09a4-731c-47db-b543-66c3282def00 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.241287] env[70013]: DEBUG oslo_vmware.api [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 956.241287] env[70013]: value = "task-4231125" [ 956.241287] env[70013]: _type = "Task" [ 956.241287] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.250867] env[70013]: DEBUG oslo_vmware.api [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.302326] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231121, 'name': ReconfigVM_Task, 'duration_secs': 0.345522} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.302585] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/39557c50-866a-4510-b840-b1a6a3e3890e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.303436] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac2c36b4-8d45-4d6a-8b5b-a1b645b0c656 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.310723] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 956.310723] env[70013]: value = "task-4231126" [ 956.310723] env[70013]: _type = "Task" [ 956.310723] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.320971] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231126, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.377359] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "81e32169-6f75-4a62-9605-9c84dae7d417" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.377643] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "81e32169-6f75-4a62-9605-9c84dae7d417" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.527744] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231123, 'name': RemoveSnapshot_Task, 'duration_secs': 0.431633} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.528094] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 956.528461] env[70013]: DEBUG nova.compute.manager [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 956.529329] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bf87d8-2dd9-4c15-8df9-d978fef0d665 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.617409] env[70013]: DEBUG nova.network.neutron [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.630118] env[70013]: DEBUG oslo_vmware.api [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190301} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.633380] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.633380] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.633380] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.635025] env[70013]: INFO nova.compute.manager [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Took 1.27 seconds to destroy the instance on the hypervisor. [ 956.635258] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 956.636254] env[70013]: DEBUG nova.compute.manager [-] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 956.636254] env[70013]: DEBUG nova.network.neutron [-] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.659435] env[70013]: ERROR nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [req-5088b04e-0b5f-4444-afcf-aa0f1b4fc257] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5088b04e-0b5f-4444-afcf-aa0f1b4fc257"}]} [ 956.659824] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.660861] env[70013]: ERROR nova.compute.manager [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Traceback (most recent call last): [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] yield [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self.set_inventory_for_provider( [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5088b04e-0b5f-4444-afcf-aa0f1b4fc257"}]} [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] During handling of the above exception, another exception occurred: [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Traceback (most recent call last): [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self._delete_instance(context, instance, bdms) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self._complete_deletion(context, instance) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self._update_resource_tracker(context, instance) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self.rt.update_usage(context, instance, instance.node) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] return f(*args, **kwargs) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self._update(context.elevated(), self.compute_nodes[nodename]) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self._update_to_placement(context, compute_node, startup) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] return attempt.get(self._wrap_exception) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] six.reraise(self.value[0], self.value[1], self.value[2]) [ 956.660861] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] raise value [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self.reportclient.update_from_provider_tree( [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] with catch_all(pd.uuid): [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] self.gen.throw(typ, value, traceback) [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] raise exception.ResourceProviderSyncFailed() [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 956.662151] env[70013]: ERROR nova.compute.manager [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] [ 956.663764] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.666s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.665929] env[70013]: INFO nova.compute.claims [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.753238] env[70013]: DEBUG oslo_vmware.api [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231125, 'name': PowerOffVM_Task, 'duration_secs': 0.200641} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.753682] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.753917] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.754309] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3017f4f-77a0-493c-a652-8c624a04412e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.821249] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231126, 'name': Rename_Task, 'duration_secs': 0.15527} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.821606] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.821906] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f634b5d-86df-4bbd-ae50-95ea4d89b5ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.829909] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 956.829909] env[70013]: value = "task-4231128" [ 956.829909] env[70013]: _type = "Task" [ 956.829909] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.836203] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.836461] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.836693] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleting the datastore file [datastore2] 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.837620] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05042011-5781-4e86-8485-947a3b1238a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.844107] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231128, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.845630] env[70013]: DEBUG oslo_vmware.api [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 956.845630] env[70013]: value = "task-4231129" [ 956.845630] env[70013]: _type = "Task" [ 956.845630] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.854527] env[70013]: DEBUG oslo_vmware.api [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.044975] env[70013]: INFO nova.compute.manager [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Shelve offloading [ 957.118629] env[70013]: INFO nova.compute.manager [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 8d28f827-c773-4f49-b89d-7d3dd8df4e70] Took 1.04 seconds to deallocate network for instance. [ 957.171791] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.800s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.343251] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231128, 'name': PowerOnVM_Task} progress is 87%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.357629] env[70013]: DEBUG nova.compute.manager [req-db0e3f4d-645d-4d2b-ab43-dc8332dd54e5 req-4700df73-79cd-4cf6-a8d9-38358c7d13ec service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-vif-deleted-6114ded0-69f6-4591-bc0f-defc91e44362 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 957.357840] env[70013]: INFO nova.compute.manager [req-db0e3f4d-645d-4d2b-ab43-dc8332dd54e5 req-4700df73-79cd-4cf6-a8d9-38358c7d13ec service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Neutron deleted interface 6114ded0-69f6-4591-bc0f-defc91e44362; detaching it from the instance and deleting it from the info cache [ 957.358184] env[70013]: DEBUG nova.network.neutron [req-db0e3f4d-645d-4d2b-ab43-dc8332dd54e5 req-4700df73-79cd-4cf6-a8d9-38358c7d13ec service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Updating instance_info_cache with network_info: [{"id": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "address": "fa:16:3e:fd:3e:39", "network": {"id": "03d9009f-f8a8-4f1c-b205-36038ddbf3f1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-959921045", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd5389a-04", "ovs_interfaceid": "0dd5389a-041f-43b0-949a-78f2655dc2ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "address": "fa:16:3e:1f:7f:fb", "network": {"id": "3c2bbab7-386e-40f5-9bf8-fdb3a1d7d0e3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1633557551", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894fb3a3-f1", "ovs_interfaceid": "894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.367028] env[70013]: DEBUG oslo_vmware.api [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254551} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.367482] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.367742] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.367985] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.368235] env[70013]: INFO nova.compute.manager [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Took 1.15 seconds to destroy the instance on the hypervisor. [ 957.368646] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 957.369681] env[70013]: DEBUG nova.compute.manager [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 957.369681] env[70013]: DEBUG nova.network.neutron [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.556017] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.556017] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f019d58-037e-4a75-8321-3eed8385c3e9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.563170] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 957.563170] env[70013]: value = "task-4231130" [ 957.563170] env[70013]: _type = "Task" [ 957.563170] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.583524] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 957.584430] env[70013]: DEBUG nova.compute.manager [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 957.585455] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85691d22-15ea-4074-888b-885b10b2b99c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.594575] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.595187] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.595187] env[70013]: DEBUG nova.network.neutron [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.721729] env[70013]: DEBUG nova.scheduler.client.report [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 957.749283] env[70013]: DEBUG nova.scheduler.client.report [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 957.749642] env[70013]: DEBUG nova.compute.provider_tree [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 957.765387] env[70013]: DEBUG nova.scheduler.client.report [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 957.793506] env[70013]: DEBUG nova.scheduler.client.report [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 957.851455] env[70013]: DEBUG oslo_vmware.api [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231128, 'name': PowerOnVM_Task, 'duration_secs': 0.885028} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.851748] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.851949] env[70013]: INFO nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Took 9.18 seconds to spawn the instance on the hypervisor. [ 957.852144] env[70013]: DEBUG nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 957.852972] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b528d4c8-fe9b-4af9-a465-62de37bfb3a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.868345] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07798843-a97f-49d4-8e7a-03fdf2e0833d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.879731] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1964627-97e5-4f6c-9755-d38d2acc6664 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.919262] env[70013]: DEBUG nova.compute.manager [req-db0e3f4d-645d-4d2b-ab43-dc8332dd54e5 req-4700df73-79cd-4cf6-a8d9-38358c7d13ec service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Detach interface failed, port_id=6114ded0-69f6-4591-bc0f-defc91e44362, reason: Instance ab091ef4-9828-438f-8e95-4370bfbd34de could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 958.160277] env[70013]: INFO nova.scheduler.client.report [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Deleted allocations for instance 8d28f827-c773-4f49-b89d-7d3dd8df4e70 [ 958.351855] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365c375a-a22e-456f-bd31-c42e59108e6a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.361096] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3c0827-f8a7-4fcd-a9d5-a3c54be7525f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.400981] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ca21cf-cb1f-4dce-be54-5779f2cb58bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.407107] env[70013]: INFO nova.compute.manager [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Took 43.60 seconds to build instance. [ 958.414544] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9319f9ce-2434-4517-8f98-047da99bdc56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.433966] env[70013]: DEBUG nova.compute.provider_tree [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.559896] env[70013]: DEBUG nova.network.neutron [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updating instance_info_cache with network_info: [{"id": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "address": "fa:16:3e:aa:b4:a5", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72c3d938-1c", "ovs_interfaceid": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.561572] env[70013]: DEBUG nova.network.neutron [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.675243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a4790199-95bf-437b-87ff-4b2a3af8bd1b tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "8d28f827-c773-4f49-b89d-7d3dd8df4e70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.473s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.701214] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.857435] env[70013]: DEBUG nova.network.neutron [-] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.909138] env[70013]: INFO nova.compute.manager [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Rescuing [ 958.909404] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.909551] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquired lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.909717] env[70013]: DEBUG nova.network.neutron [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.911132] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eb4feab0-90e6-476d-ab4f-499ef9b50755 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "39557c50-866a-4510-b840-b1a6a3e3890e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.466s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.940108] env[70013]: DEBUG nova.scheduler.client.report [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.064032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.066159] env[70013]: INFO nova.compute.manager [-] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Took 1.70 seconds to deallocate network for instance. [ 959.178293] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 959.361976] env[70013]: INFO nova.compute.manager [-] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Took 2.72 seconds to deallocate network for instance. [ 959.392077] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 959.392994] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0648f6-8a23-4ed9-b85b-a4008556a420 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.403867] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.403956] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47b8de26-4e24-4f90-9a74-f788d6e3fc31 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.417459] env[70013]: DEBUG nova.compute.manager [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 959.444580] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.781s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.448073] env[70013]: DEBUG nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 959.456022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 22.225s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.484181] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.484181] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.485107] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleting the datastore file [datastore2] 30fcb8c7-0947-426e-9aa5-081446e31c28 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.485107] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a673fd01-91c1-48a1-bc1a-6ab8a255c021 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.492622] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 959.492622] env[70013]: value = "task-4231132" [ 959.492622] env[70013]: _type = "Task" [ 959.492622] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.506220] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.577269] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.632643] env[70013]: DEBUG nova.network.neutron [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.707983] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.778424] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Acquiring lock "a8bb5ed8-00cc-493d-9ab4-264d878d4135" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.778677] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Lock "a8bb5ed8-00cc-493d-9ab4-264d878d4135" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.797822] env[70013]: DEBUG nova.compute.manager [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-vif-deleted-0dd5389a-041f-43b0-949a-78f2655dc2ee {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 959.798022] env[70013]: DEBUG nova.compute.manager [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Received event network-vif-deleted-f6f11372-627a-4dc6-8fe9-e20fcc0a36ed {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 959.798185] env[70013]: DEBUG nova.compute.manager [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Received event network-vif-deleted-894fb3a3-f1d1-438f-abc0-6fa78a9e8b4a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 959.798391] env[70013]: DEBUG nova.compute.manager [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Received event network-vif-unplugged-72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 959.798626] env[70013]: DEBUG oslo_concurrency.lockutils [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] Acquiring lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.798841] env[70013]: DEBUG oslo_concurrency.lockutils [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.799022] env[70013]: DEBUG oslo_concurrency.lockutils [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.799193] env[70013]: DEBUG nova.compute.manager [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] No waiting events found dispatching network-vif-unplugged-72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 959.799361] env[70013]: WARNING nova.compute.manager [req-1b679b4a-e983-4a39-b571-7340138db4ec req-7d928b91-874d-4a85-9708-2ca245c49522 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Received unexpected event network-vif-unplugged-72c3d938-1cba-4e01-8176-aa8e8a245c4c for instance with vm_state shelved and task_state shelving_offloading. [ 959.870143] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.938795] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.969756] env[70013]: DEBUG nova.compute.utils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 959.971474] env[70013]: DEBUG nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 959.972017] env[70013]: DEBUG nova.network.neutron [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 960.004165] env[70013]: DEBUG oslo_vmware.api [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231132, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149653} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.004477] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.004674] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 960.004932] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 960.026586] env[70013]: DEBUG nova.policy [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd357d73286774d2cbe0ab68494c7c57b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75ec7d6e9ea54c06bd3b7605f16b391a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 960.029530] env[70013]: INFO nova.scheduler.client.report [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted allocations for instance 30fcb8c7-0947-426e-9aa5-081446e31c28 [ 960.138021] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Releasing lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 960.347175] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "5c855194-feef-403f-9279-a32d5492c353" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 960.347175] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "5c855194-feef-403f-9279-a32d5492c353" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.475051] env[70013]: DEBUG nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 960.482089] env[70013]: DEBUG nova.network.neutron [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Successfully created port: 4edd213b-3ed8-4731-8128-4caf383b2277 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.533933] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 960.869877] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "30fcb8c7-0947-426e-9aa5-081446e31c28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.492937] env[70013]: DEBUG nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 28a56a75-9b30-4121-8252-a9e57287441c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 63c3b57c-022a-4eee-b215-6dd89142e659 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance fc473d40-b57b-437e-9511-58a0ba700a69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 09544922-6dc4-48b5-8cfd-e91e7f74c13f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2c0e4326-c33a-42bb-b793-a100157b1c03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 85488f5e-bd8a-4bcf-8420-744c54176c2e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 67b90fec-a6fd-471d-a425-82bc63de572a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance eab0c393-4d3d-4659-8225-c82ddadb330b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.511754] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.512205] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 965480de-4fe2-48ce-94ad-0838d78239dd is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 961.512205] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 472b8218-51c3-492e-96ff-5fa99df4cc4c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.512205] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 961.512290] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ab091ef4-9828-438f-8e95-4370bfbd34de is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 961.512403] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.512638] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.512638] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8dbef727-13e9-4231-95a8-65d015ee13be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 961.532106] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f3568bb8fb91151f56a35e7456aeb11a',container_format='bare',created_at=2025-05-19T11:36:42Z,direct_url=,disk_format='vmdk',id=ca4066da-1b0a-465f-b38a-e1e6b704308d,min_disk=1,min_ram=0,name='tempest-test-snap-686289591',owner='75ec7d6e9ea54c06bd3b7605f16b391a',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-05-19T11:36:57Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 961.532394] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.532584] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 961.532809] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.533023] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 961.533199] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 961.533437] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 961.533615] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 961.533806] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 961.533975] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 961.534193] env[70013]: DEBUG nova.virt.hardware [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 961.536110] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d3297c-b8a2-4949-85cd-10dce842b902 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.547552] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11820f91-a91c-4b91-bd0d-41576feaecbb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.678485] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.678926] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55f3bcc6-f2ec-4a89-bf21-29651f2da64b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.686651] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 961.686651] env[70013]: value = "task-4231133" [ 961.686651] env[70013]: _type = "Task" [ 961.686651] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.699697] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231133, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.903886] env[70013]: DEBUG nova.compute.manager [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Received event network-changed-72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 961.904197] env[70013]: DEBUG nova.compute.manager [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Refreshing instance network info cache due to event network-changed-72c3d938-1cba-4e01-8176-aa8e8a245c4c. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 961.904426] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] Acquiring lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.904634] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] Acquired lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.904841] env[70013]: DEBUG nova.network.neutron [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Refreshing network info cache for port 72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.016895] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 0ea84d0d-905c-428d-8abb-2781c817f08f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 962.082112] env[70013]: DEBUG nova.network.neutron [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Successfully updated port: 4edd213b-3ed8-4731-8128-4caf383b2277 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.197513] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231133, 'name': PowerOffVM_Task, 'duration_secs': 0.200003} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.197835] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.198777] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08388585-5ca2-4c4f-a2ab-d0b0c92fa962 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.219331] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9aca6f-4ca2-488d-95aa-56823ddc4e13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.253599] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.253902] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb7086c8-0c51-45ca-b15c-36a048901978 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.261864] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 962.261864] env[70013]: value = "task-4231134" [ 962.261864] env[70013]: _type = "Task" [ 962.261864] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.270871] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231134, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.520657] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 54d39dfd-6429-4009-8d26-22c1fa46243e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 962.584952] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.585158] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.585524] env[70013]: DEBUG nova.network.neutron [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.602885] env[70013]: DEBUG nova.network.neutron [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updated VIF entry in instance network info cache for port 72c3d938-1cba-4e01-8176-aa8e8a245c4c. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.603244] env[70013]: DEBUG nova.network.neutron [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updating instance_info_cache with network_info: [{"id": "72c3d938-1cba-4e01-8176-aa8e8a245c4c", "address": "fa:16:3e:aa:b4:a5", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": null, "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap72c3d938-1c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.773391] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 962.773623] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.773882] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.774046] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.774232] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.774510] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed28d448-dd59-470a-b076-df4e85cd9d83 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.785642] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.785846] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.786675] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1da7c59-d4e2-414b-8efa-af4b8c5e973b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.792627] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 962.792627] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529a9374-9e6a-cf9e-2985-f92395c61539" [ 962.792627] env[70013]: _type = "Task" [ 962.792627] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.801373] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529a9374-9e6a-cf9e-2985-f92395c61539, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.023723] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 963.106186] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff35683d-6b4b-4776-b87f-381bc95a0c5d req-44a53ad4-7984-40e6-a505-ed8b4de3cbd1 service nova] Releasing lock "refresh_cache-30fcb8c7-0947-426e-9aa5-081446e31c28" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.117819] env[70013]: DEBUG nova.network.neutron [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 963.174035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.174035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.174035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.175028] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.175028] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 963.176624] env[70013]: INFO nova.compute.manager [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Terminating instance [ 963.252737] env[70013]: DEBUG nova.network.neutron [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Updating instance_info_cache with network_info: [{"id": "4edd213b-3ed8-4731-8128-4caf383b2277", "address": "fa:16:3e:5a:d6:a1", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4edd213b-3e", "ovs_interfaceid": "4edd213b-3ed8-4731-8128-4caf383b2277", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.303048] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529a9374-9e6a-cf9e-2985-f92395c61539, 'name': SearchDatastore_Task, 'duration_secs': 0.010362} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.303847] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8414e441-0df4-4056-b6e6-ff6d2f5b2db6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.309629] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 963.309629] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52174cda-7707-6123-a6d2-a52c04f1a6ad" [ 963.309629] env[70013]: _type = "Task" [ 963.309629] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.317668] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52174cda-7707-6123-a6d2-a52c04f1a6ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.526896] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 963.680595] env[70013]: DEBUG nova.compute.manager [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 963.680856] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.681874] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeef5aba-ddfe-4366-b48d-5c831e1c6db0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.689904] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.690174] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1346af9-6326-460c-910b-510106276a0d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.696803] env[70013]: DEBUG oslo_vmware.api [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 963.696803] env[70013]: value = "task-4231135" [ 963.696803] env[70013]: _type = "Task" [ 963.696803] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.706365] env[70013]: DEBUG oslo_vmware.api [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.756325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.756770] env[70013]: DEBUG nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance network_info: |[{"id": "4edd213b-3ed8-4731-8128-4caf383b2277", "address": "fa:16:3e:5a:d6:a1", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4edd213b-3e", "ovs_interfaceid": "4edd213b-3ed8-4731-8128-4caf383b2277", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 963.757344] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:d6:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4edd213b-3ed8-4731-8128-4caf383b2277', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 963.765565] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 963.765813] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 963.766061] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6cf4c338-02a3-4f37-b491-997962a3df15 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.786941] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 963.786941] env[70013]: value = "task-4231136" [ 963.786941] env[70013]: _type = "Task" [ 963.786941] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.796157] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231136, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.821615] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52174cda-7707-6123-a6d2-a52c04f1a6ad, 'name': SearchDatastore_Task, 'duration_secs': 0.01005} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.821947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.822233] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. {{(pid=70013) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 963.822533] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ca4a720-d843-4397-a304-f38498e90710 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.828974] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 963.828974] env[70013]: value = "task-4231137" [ 963.828974] env[70013]: _type = "Task" [ 963.828974] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.837466] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231137, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.026030] env[70013]: DEBUG nova.compute.manager [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Received event network-vif-plugged-4edd213b-3ed8-4731-8128-4caf383b2277 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 964.026329] env[70013]: DEBUG oslo_concurrency.lockutils [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] Acquiring lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.026628] env[70013]: DEBUG oslo_concurrency.lockutils [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] Lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.026839] env[70013]: DEBUG oslo_concurrency.lockutils [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] Lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.027062] env[70013]: DEBUG nova.compute.manager [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] No waiting events found dispatching network-vif-plugged-4edd213b-3ed8-4731-8128-4caf383b2277 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 964.027286] env[70013]: WARNING nova.compute.manager [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Received unexpected event network-vif-plugged-4edd213b-3ed8-4731-8128-4caf383b2277 for instance with vm_state building and task_state spawning. [ 964.027491] env[70013]: DEBUG nova.compute.manager [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Received event network-changed-4edd213b-3ed8-4731-8128-4caf383b2277 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 964.027679] env[70013]: DEBUG nova.compute.manager [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Refreshing instance network info cache due to event network-changed-4edd213b-3ed8-4731-8128-4caf383b2277. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 964.027909] env[70013]: DEBUG oslo_concurrency.lockutils [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] Acquiring lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.028121] env[70013]: DEBUG oslo_concurrency.lockutils [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] Acquired lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.028355] env[70013]: DEBUG nova.network.neutron [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Refreshing network info cache for port 4edd213b-3ed8-4731-8128-4caf383b2277 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.030489] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance b292a985-3d2a-4cf3-a09b-8e72d8d21078 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 964.210043] env[70013]: DEBUG oslo_vmware.api [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231135, 'name': PowerOffVM_Task, 'duration_secs': 0.229265} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.210367] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.210624] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.211014] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1c85ae9-f489-4de2-8524-b8c0bfc63e58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.299489] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231136, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.313634] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.313919] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.314125] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Deleting the datastore file [datastore1] ef64a05a-b514-4c35-81d3-664ae1ad3ff1 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.314421] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-794704f4-c85c-4375-b848-8608e0c99bf6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.322047] env[70013]: DEBUG oslo_vmware.api [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 964.322047] env[70013]: value = "task-4231139" [ 964.322047] env[70013]: _type = "Task" [ 964.322047] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.331479] env[70013]: DEBUG oslo_vmware.api [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.342815] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231137, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.535479] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2070989d-6c90-4eb1-8508-7587ed9659d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 964.735866] env[70013]: DEBUG nova.network.neutron [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Updated VIF entry in instance network info cache for port 4edd213b-3ed8-4731-8128-4caf383b2277. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 964.736244] env[70013]: DEBUG nova.network.neutron [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Updating instance_info_cache with network_info: [{"id": "4edd213b-3ed8-4731-8128-4caf383b2277", "address": "fa:16:3e:5a:d6:a1", "network": {"id": "46c185a4-fd00-42fe-a6e6-b546a0baa1f7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-331737986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75ec7d6e9ea54c06bd3b7605f16b391a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4edd213b-3e", "ovs_interfaceid": "4edd213b-3ed8-4731-8128-4caf383b2277", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.797844] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231136, 'name': CreateVM_Task, 'duration_secs': 0.581839} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.798037] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 964.798740] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.798909] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.799297] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 964.799571] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6203d0db-1fe7-4f92-94fd-19f5c071bd55 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.805029] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 964.805029] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52158283-af67-e308-6fc2-63c526fb683b" [ 964.805029] env[70013]: _type = "Task" [ 964.805029] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.814202] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52158283-af67-e308-6fc2-63c526fb683b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.832371] env[70013]: DEBUG oslo_vmware.api [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180291} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.835511] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.835784] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 964.835979] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 964.836173] env[70013]: INFO nova.compute.manager [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Took 1.16 seconds to destroy the instance on the hypervisor. [ 964.836407] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 964.836603] env[70013]: DEBUG nova.compute.manager [-] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 964.836700] env[70013]: DEBUG nova.network.neutron [-] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 964.843957] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231137, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519032} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.844217] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. [ 964.844986] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df02766e-f0b7-4efe-b819-936288e5e921 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.870022] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.870359] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a0600df-386b-4acc-bb66-2766b5ab2d6d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.890398] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 964.890398] env[70013]: value = "task-4231140" [ 964.890398] env[70013]: _type = "Task" [ 964.890398] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.905650] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231140, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.040305] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 69f67e8d-2da0-474c-92db-8388c98b61f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 965.239747] env[70013]: DEBUG oslo_concurrency.lockutils [req-f16ceba4-d3e6-4723-adc0-7bfade0a9808 req-7a23aef6-d95f-4d98-8a27-244aa55f0d26 service nova] Releasing lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 965.316404] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 965.316699] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Processing image ca4066da-1b0a-465f-b38a-e1e6b704308d {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 965.316953] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d/ca4066da-1b0a-465f-b38a-e1e6b704308d.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.317118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d/ca4066da-1b0a-465f-b38a-e1e6b704308d.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 965.317298] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.317568] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-deb581bf-16d4-4b48-9592-4e3debbd45f5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.328110] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.328310] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 965.329107] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df9dd0f0-98de-4fbf-9db6-836b2cb88278 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.335486] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 965.335486] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fadaae-c5b1-328a-c758-68ccfcd7bb0a" [ 965.335486] env[70013]: _type = "Task" [ 965.335486] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.345719] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fadaae-c5b1-328a-c758-68ccfcd7bb0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.402057] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231140, 'name': ReconfigVM_Task, 'duration_secs': 0.337397} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.402057] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.402668] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36d24ef-6290-45e6-b0ca-53ab55122beb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.430329] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-915b28c1-6345-469f-9d22-766daa9b0687 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.445918] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 965.445918] env[70013]: value = "task-4231141" [ 965.445918] env[70013]: _type = "Task" [ 965.445918] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.454796] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231141, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.544454] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 5a535178-36b5-4015-9cd4-1a3b9b7517c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 965.590605] env[70013]: DEBUG nova.network.neutron [-] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.846023] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Preparing fetch location {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 965.846555] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Fetch image to [datastore2] OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e/OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e.vmdk {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 965.846883] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Downloading stream optimized image ca4066da-1b0a-465f-b38a-e1e6b704308d to [datastore2] OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e/OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e.vmdk on the data store datastore2 as vApp {{(pid=70013) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 965.847218] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Downloading image file data ca4066da-1b0a-465f-b38a-e1e6b704308d to the ESX as VM named 'OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e' {{(pid=70013) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 965.922240] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 965.922240] env[70013]: value = "resgroup-9" [ 965.922240] env[70013]: _type = "ResourcePool" [ 965.922240] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 965.922565] env[70013]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a9cc439c-9cbc-4986-8f98-6437505c8649 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.944732] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease: (returnval){ [ 965.944732] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5ec6-31fd-da2e-3fa1-2a8f05deb44e" [ 965.944732] env[70013]: _type = "HttpNfcLease" [ 965.944732] env[70013]: } obtained for vApp import into resource pool (val){ [ 965.944732] env[70013]: value = "resgroup-9" [ 965.944732] env[70013]: _type = "ResourcePool" [ 965.944732] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 965.945095] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the lease: (returnval){ [ 965.945095] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5ec6-31fd-da2e-3fa1-2a8f05deb44e" [ 965.945095] env[70013]: _type = "HttpNfcLease" [ 965.945095] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 965.955403] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 965.955403] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5ec6-31fd-da2e-3fa1-2a8f05deb44e" [ 965.955403] env[70013]: _type = "HttpNfcLease" [ 965.955403] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 965.958626] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231141, 'name': ReconfigVM_Task, 'duration_secs': 0.182634} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.958874] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.959146] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24213595-2472-41dc-a722-bbaef85f7ebd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.966669] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 965.966669] env[70013]: value = "task-4231143" [ 965.966669] env[70013]: _type = "Task" [ 965.966669] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.977474] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231143, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.048688] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 81e32169-6f75-4a62-9605-9c84dae7d417 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 966.053248] env[70013]: DEBUG nova.compute.manager [req-da2964a3-fe72-4226-9484-f9aa37be4556 req-a9c99816-7309-4670-bfe3-4f781ec65aa8 service nova] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Received event network-vif-deleted-25ae3868-dd45-4f4c-8b91-809dcb235f5c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 966.093046] env[70013]: INFO nova.compute.manager [-] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Took 1.26 seconds to deallocate network for instance. [ 966.454251] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 966.454251] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5ec6-31fd-da2e-3fa1-2a8f05deb44e" [ 966.454251] env[70013]: _type = "HttpNfcLease" [ 966.454251] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 966.454568] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 966.454568] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b5ec6-31fd-da2e-3fa1-2a8f05deb44e" [ 966.454568] env[70013]: _type = "HttpNfcLease" [ 966.454568] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 966.455323] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af75b3e9-91a5-440b-9ca6-57196485b315 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.463285] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52680887-5ebe-d477-8f54-776b6cf59b58/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 966.463568] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52680887-5ebe-d477-8f54-776b6cf59b58/disk-0.vmdk. {{(pid=70013) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 966.532752] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b93c1b14-4756-40ff-8865-6dfebd469f1b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.534589] env[70013]: DEBUG oslo_vmware.api [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231143, 'name': PowerOnVM_Task, 'duration_secs': 0.42901} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.536179] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.539914] env[70013]: DEBUG nova.compute.manager [None req-dc4e6511-80b8-4706-9365-084bea96d486 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 966.541161] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f13487-a228-4923-a4f1-3f7f7867bbf0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.551889] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance a8bb5ed8-00cc-493d-9ab4-264d878d4135 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 966.601498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.057613] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 5c855194-feef-403f-9279-a32d5492c353 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 967.058013] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 17 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 967.058278] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3776MB phys_disk=200GB used_disk=17GB total_vcpus=48 used_vcpus=17 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 967.239798] env[70013]: DEBUG nova.compute.manager [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 967.240009] env[70013]: DEBUG nova.compute.manager [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing instance network info cache due to event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 967.240245] env[70013]: DEBUG oslo_concurrency.lockutils [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] Acquiring lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.240389] env[70013]: DEBUG oslo_concurrency.lockutils [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] Acquired lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 967.240635] env[70013]: DEBUG nova.network.neutron [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.361994] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Completed reading data from the image iterator. {{(pid=70013) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 967.362240] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52680887-5ebe-d477-8f54-776b6cf59b58/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 967.363175] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c91b8f2-9ffe-4500-a542-0752deed572c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.372985] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52680887-5ebe-d477-8f54-776b6cf59b58/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 967.373181] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52680887-5ebe-d477-8f54-776b6cf59b58/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 967.375291] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-dd27e4a7-55f9-4f7f-a4ca-6c447edca6e9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.506038] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074143e6-0456-4a2e-9989-c1326c8a98b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.514226] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ceab27-c920-4614-aeaf-c38304d3dd05 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.549120] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052bf7f9-ea49-4289-9737-e6aabdff9bff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.558767] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b068d451-2c67-4c7b-8b63-7b1d12a51985 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.574533] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.578191] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52680887-5ebe-d477-8f54-776b6cf59b58/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 967.578441] env[70013]: INFO nova.virt.vmwareapi.images [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Downloaded image file data ca4066da-1b0a-465f-b38a-e1e6b704308d [ 967.579733] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c2007c-4a1c-46cc-9675-097166b21921 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.597194] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d86af8a7-623e-4f33-bba2-ef2e58110b7f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.628964] env[70013]: INFO nova.virt.vmwareapi.images [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] The imported VM was unregistered [ 967.631499] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Caching image {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 967.631744] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Creating directory with path [datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.632680] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b6a6cbb-1fd7-402d-9882-bd7e7046df20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.644135] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Created directory with path [datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.644352] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e/OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e.vmdk to [datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d/ca4066da-1b0a-465f-b38a-e1e6b704308d.vmdk. {{(pid=70013) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 967.644642] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-3300bebc-ade0-4c84-b30b-d9634c1bb4a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.654539] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 967.654539] env[70013]: value = "task-4231145" [ 967.654539] env[70013]: _type = "Task" [ 967.654539] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.662403] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231145, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.969088] env[70013]: DEBUG nova.network.neutron [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updated VIF entry in instance network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.969481] env[70013]: DEBUG nova.network.neutron [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.083146] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 968.164347] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231145, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.473204] env[70013]: DEBUG oslo_concurrency.lockutils [req-ee58d9ed-ff45-4c62-8c49-5ff77e574bc7 req-40408901-f645-47e2-a60b-d92249a4ef62 service nova] Releasing lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 968.587598] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 968.587962] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.134s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 968.588324] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.765s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 968.588634] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 968.591050] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.268s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 968.593010] env[70013]: INFO nova.compute.claims [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 968.596326] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.596553] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Cleaning up deleted instances {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11864}} [ 968.623487] env[70013]: INFO nova.scheduler.client.report [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Deleted allocations for instance 67b90fec-a6fd-471d-a425-82bc63de572a [ 968.670896] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231145, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.110032] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] There are 32 instances to clean {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11873}} [ 969.110384] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: d84d8426-b3fd-4f0d-8410-0ee24fedb35d] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 969.131709] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3b141172-355e-4acc-ba6d-f413ab7c2de9 tempest-SecurityGroupsTestJSON-1428022973 tempest-SecurityGroupsTestJSON-1428022973-project-member] Lock "67b90fec-a6fd-471d-a425-82bc63de572a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.977s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.169063] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231145, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.618260] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 0a56e30f-fb07-4a5e-9d69-9603fd38ff9f] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 969.626484] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e592b737-6756-48e3-a17b-cb23329a5b08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.636032] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b45d66-4a7f-46c6-89e9-e8abd7fb8371 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.673559] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616f6cc6-086b-42cb-b537-f7e5a5e7e065 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.687436] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a7f564-637e-4e94-a368-09c303073915 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.691803] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231145, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.706680] env[70013]: DEBUG nova.compute.provider_tree [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.781186] env[70013]: DEBUG nova.compute.manager [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 969.781413] env[70013]: DEBUG nova.compute.manager [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing instance network info cache due to event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 969.781766] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] Acquiring lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.781894] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] Acquired lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 969.782066] env[70013]: DEBUG nova.network.neutron [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.123201] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: a53389ce-d0c2-48ae-83e3-ee58fbdbfb01] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 970.179624] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231145, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.209740] env[70013]: DEBUG nova.scheduler.client.report [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 970.571481] env[70013]: DEBUG nova.network.neutron [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updated VIF entry in instance network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.571894] env[70013]: DEBUG nova.network.neutron [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.626438] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 0b0b2769-74ba-4ba2-9e8b-148cc2976321] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 970.682087] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231145, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.576195} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.682087] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e/OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e.vmdk to [datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d/ca4066da-1b0a-465f-b38a-e1e6b704308d.vmdk. [ 970.682290] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Cleaning up location [datastore2] OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 970.682429] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_65d227b1-61ac-4a12-a2d3-d6002a6f8f4e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.682697] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6df9f6a-dc42-468e-9716-a61ea132997e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.693428] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 970.693428] env[70013]: value = "task-4231146" [ 970.693428] env[70013]: _type = "Task" [ 970.693428] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.703390] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231146, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.717820] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.126s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.718162] env[70013]: DEBUG nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 970.723131] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.351s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.724227] env[70013]: INFO nova.compute.claims [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.075020] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b1c5156-3ab4-40da-9355-7cd596720b3e req-d01d5c25-247d-4969-9bc2-c7d628393b1b service nova] Releasing lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.134568] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 6b458a7e-f786-41f6-b8ac-9f6f4ddcb268] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 971.205013] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231146, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040402} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.207197] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.207589] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d/ca4066da-1b0a-465f-b38a-e1e6b704308d.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.207879] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d/ca4066da-1b0a-465f-b38a-e1e6b704308d.vmdk to [datastore2] 8dbef727-13e9-4231-95a8-65d015ee13be/8dbef727-13e9-4231-95a8-65d015ee13be.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 971.208528] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-269675cd-d1a6-402b-ba93-0e917131ee14 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.216953] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 971.216953] env[70013]: value = "task-4231147" [ 971.216953] env[70013]: _type = "Task" [ 971.216953] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.226119] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231147, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.230724] env[70013]: DEBUG nova.compute.utils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 971.232838] env[70013]: DEBUG nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 971.233114] env[70013]: DEBUG nova.network.neutron [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 971.307075] env[70013]: DEBUG nova.policy [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88d1be071a1c47d8b7664380ae025351', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68ad346a8e3a465da70d3a7de825ac6d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 971.641746] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: be9a565b-c74c-442e-8087-cc4579b1c8e7] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 971.734783] env[70013]: DEBUG nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 971.738603] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231147, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.875028] env[70013]: DEBUG nova.compute.manager [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 971.875514] env[70013]: DEBUG nova.compute.manager [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing instance network info cache due to event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 971.875765] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] Acquiring lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.875996] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] Acquired lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.876192] env[70013]: DEBUG nova.network.neutron [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.020445] env[70013]: DEBUG nova.network.neutron [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Successfully created port: d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 972.148076] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: f715cf45-81f1-4979-abfc-455b3e64f00e] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 972.167161] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "39557c50-866a-4510-b840-b1a6a3e3890e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.167274] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "39557c50-866a-4510-b840-b1a6a3e3890e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.167431] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "39557c50-866a-4510-b840-b1a6a3e3890e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.167622] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "39557c50-866a-4510-b840-b1a6a3e3890e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.167815] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "39557c50-866a-4510-b840-b1a6a3e3890e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.170704] env[70013]: INFO nova.compute.manager [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Terminating instance [ 972.229938] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231147, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.316289] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f090ca3b-715d-4a20-b71d-7ba9f6ca4dff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.326225] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d477da74-13fc-4a80-a4a9-de3ee7a2d703 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.390390] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3987bc-0ceb-436e-80c4-e58e5d6bd803 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.403540] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda46ab6-528e-4fe4-ae4d-1e8318c49627 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.427992] env[70013]: DEBUG nova.compute.provider_tree [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.656836] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: b1d844cc-dcdc-47e2-85c3-484d0862be60] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 972.679628] env[70013]: DEBUG nova.compute.manager [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 972.680010] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.681180] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7dd921-6a2a-4dcd-b764-d76b1425cb16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.687490] env[70013]: DEBUG nova.network.neutron [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updated VIF entry in instance network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.687887] env[70013]: DEBUG nova.network.neutron [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.691917] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.692574] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d9e8937-f7a7-4e33-8932-c74a930ab200 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.702448] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 972.702448] env[70013]: value = "task-4231148" [ 972.702448] env[70013]: _type = "Task" [ 972.702448] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.714686] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.728938] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231147, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.747658] env[70013]: DEBUG nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 972.784620] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 972.784945] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.785213] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 972.785421] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.785578] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 972.785730] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 972.785991] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 972.786198] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 972.786405] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 972.786681] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 972.786841] env[70013]: DEBUG nova.virt.hardware [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 972.787935] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70054e3-26b2-4be0-8ae7-9a86f2782945 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.798962] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9032d417-87d8-42c6-bc9f-27b4859e4e7e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.935018] env[70013]: DEBUG nova.scheduler.client.report [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 973.161737] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 02ef7875-7ec9-4409-aaa5-71ed669f3780] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 973.193968] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] Releasing lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 973.194366] env[70013]: DEBUG nova.compute.manager [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 973.194760] env[70013]: DEBUG nova.compute.manager [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing instance network info cache due to event network-changed-7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 973.194834] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] Acquiring lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.194935] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] Acquired lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 973.195118] env[70013]: DEBUG nova.network.neutron [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Refreshing network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.217639] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231148, 'name': PowerOffVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.232643] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231147, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.441393] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.719s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.441939] env[70013]: DEBUG nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 973.444894] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.720s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.445193] env[70013]: DEBUG nova.objects.instance [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 973.611747] env[70013]: DEBUG nova.compute.manager [req-dd422e9f-a017-4653-a8c5-4914c42a05ea req-edfbece6-4bb4-44da-a8e5-740a52e79144 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received event network-vif-plugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 973.612245] env[70013]: DEBUG oslo_concurrency.lockutils [req-dd422e9f-a017-4653-a8c5-4914c42a05ea req-edfbece6-4bb4-44da-a8e5-740a52e79144 service nova] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.614277] env[70013]: DEBUG oslo_concurrency.lockutils [req-dd422e9f-a017-4653-a8c5-4914c42a05ea req-edfbece6-4bb4-44da-a8e5-740a52e79144 service nova] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.614277] env[70013]: DEBUG oslo_concurrency.lockutils [req-dd422e9f-a017-4653-a8c5-4914c42a05ea req-edfbece6-4bb4-44da-a8e5-740a52e79144 service nova] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.614277] env[70013]: DEBUG nova.compute.manager [req-dd422e9f-a017-4653-a8c5-4914c42a05ea req-edfbece6-4bb4-44da-a8e5-740a52e79144 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] No waiting events found dispatching network-vif-plugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 973.614277] env[70013]: WARNING nova.compute.manager [req-dd422e9f-a017-4653-a8c5-4914c42a05ea req-edfbece6-4bb4-44da-a8e5-740a52e79144 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received unexpected event network-vif-plugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 for instance with vm_state building and task_state spawning. [ 973.651157] env[70013]: DEBUG nova.network.neutron [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Successfully updated port: d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 973.667101] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 352790fe-057f-45ee-aff0-549b456fd181] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 973.719076] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231148, 'name': PowerOffVM_Task, 'duration_secs': 0.540625} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.719883] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.720220] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 973.723956] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9c4693a-4cd7-4482-ac27-3c8d1ba9215e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.733185] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231147, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.820209] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 973.820446] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 973.821164] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Deleting the datastore file [datastore2] 39557c50-866a-4510-b840-b1a6a3e3890e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.821164] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48234444-8efe-4423-a4a1-671a6193f382 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.828912] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for the task: (returnval){ [ 973.828912] env[70013]: value = "task-4231150" [ 973.828912] env[70013]: _type = "Task" [ 973.828912] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.837717] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.940457] env[70013]: DEBUG nova.network.neutron [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updated VIF entry in instance network info cache for port 7beab378-23cf-4f47-8004-b917e0a1e482. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 973.941557] env[70013]: DEBUG nova.network.neutron [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [{"id": "7beab378-23cf-4f47-8004-b917e0a1e482", "address": "fa:16:3e:33:6d:97", "network": {"id": "058139da-4d9c-40a1-9e9d-5eaa273f9907", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1040886485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7ca0b60f856f4cd7a243629fcee1cd42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7beab378-23", "ovs_interfaceid": "7beab378-23cf-4f47-8004-b917e0a1e482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.950748] env[70013]: DEBUG nova.compute.utils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 973.957799] env[70013]: DEBUG nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 973.957799] env[70013]: DEBUG nova.network.neutron [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 974.001783] env[70013]: DEBUG nova.policy [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2601a89c7a09474bb4da0b707d9987f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3c45f05e8334dcca93a5c6d4ef4c68b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 974.154408] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.154408] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 974.154408] env[70013]: DEBUG nova.network.neutron [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 974.172787] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 5c5fbdae-4916-4beb-ada0-57f36fb0f84c] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 974.234149] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231147, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.598052} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.234420] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ca4066da-1b0a-465f-b38a-e1e6b704308d/ca4066da-1b0a-465f-b38a-e1e6b704308d.vmdk to [datastore2] 8dbef727-13e9-4231-95a8-65d015ee13be/8dbef727-13e9-4231-95a8-65d015ee13be.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.235297] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d6e6df-b5e7-4209-840e-caeb4942997d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.260144] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 8dbef727-13e9-4231-95a8-65d015ee13be/8dbef727-13e9-4231-95a8-65d015ee13be.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.260484] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-468cb756-e8cb-44f8-8ac8-823c64adc18a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.280620] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.280883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.286219] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 974.286219] env[70013]: value = "task-4231151" [ 974.286219] env[70013]: _type = "Task" [ 974.286219] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.298922] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231151, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.339092] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.357593] env[70013]: DEBUG nova.network.neutron [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Successfully created port: 20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.445275] env[70013]: DEBUG oslo_concurrency.lockutils [req-1f1a7d89-1756-4618-b7f3-df72efcba2b3 req-04e5188d-627f-4bf1-90ff-57574bff7bec service nova] Releasing lock "refresh_cache-39557c50-866a-4510-b840-b1a6a3e3890e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.457970] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e91c140d-d0b8-4fa5-b5c6-94a47d888448 tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.459442] env[70013]: DEBUG nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 974.462565] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.099s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.464038] env[70013]: INFO nova.compute.claims [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.678580] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: ac8aa41c-375a-4231-9a5d-12d6c4bba44c] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 974.689683] env[70013]: DEBUG nova.network.neutron [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 974.796645] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231151, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.842437] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.843124] env[70013]: DEBUG nova.network.neutron [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.181693] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 77c12460-9cfa-41c9-a210-238a470d9ccd] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 975.298582] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231151, 'name': ReconfigVM_Task, 'duration_secs': 0.735466} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.298998] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 8dbef727-13e9-4231-95a8-65d015ee13be/8dbef727-13e9-4231-95a8-65d015ee13be.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.299686] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e76cd39-e0ab-4fd0-9cd8-f9a71e6fdebe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.308698] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 975.308698] env[70013]: value = "task-4231152" [ 975.308698] env[70013]: _type = "Task" [ 975.308698] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.318820] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231152, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.343272] env[70013]: DEBUG oslo_vmware.api [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Task: {'id': task-4231150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.06177} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.343559] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.343756] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.343935] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.344158] env[70013]: INFO nova.compute.manager [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Took 2.66 seconds to destroy the instance on the hypervisor. [ 975.344648] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 975.344819] env[70013]: DEBUG nova.compute.manager [-] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 975.345081] env[70013]: DEBUG nova.network.neutron [-] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 975.347111] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 975.347470] env[70013]: DEBUG nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Instance network_info: |[{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 975.348141] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:be:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8aacd8d-2fc2-477b-9d82-e101e20b0132', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.356865] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating folder: Project (68ad346a8e3a465da70d3a7de825ac6d). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 975.357576] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ce21e4f-c962-4419-b1f7-27ad39fb7da3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.371173] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Created folder: Project (68ad346a8e3a465da70d3a7de825ac6d) in parent group-v836999. [ 975.371432] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating folder: Instances. Parent ref: group-v837163. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 975.372104] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa90af2d-3321-4864-8109-039a9a888091 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.384716] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Created folder: Instances in parent group-v837163. [ 975.385422] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 975.385422] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 975.385557] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5900a406-cd0f-4761-8a03-d2b7ea35037e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.409622] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 975.409622] env[70013]: value = "task-4231155" [ 975.409622] env[70013]: _type = "Task" [ 975.409622] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.426025] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231155, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.478762] env[70013]: DEBUG nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 975.522468] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 975.522923] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.523176] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 975.523596] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.523822] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 975.524041] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 975.524511] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 975.524778] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 975.525015] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 975.525255] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 975.525499] env[70013]: DEBUG nova.virt.hardware [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 975.526719] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9fc719-938a-4522-9291-3e5d885493d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.536752] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbf3128-1544-4124-87fe-193852b44b1b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.650217] env[70013]: DEBUG nova.compute.manager [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received event network-changed-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 975.650469] env[70013]: DEBUG nova.compute.manager [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Refreshing instance network info cache due to event network-changed-d8aacd8d-2fc2-477b-9d82-e101e20b0132. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 975.650827] env[70013]: DEBUG oslo_concurrency.lockutils [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] Acquiring lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.650935] env[70013]: DEBUG oslo_concurrency.lockutils [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] Acquired lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.651175] env[70013]: DEBUG nova.network.neutron [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Refreshing network info cache for port d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 975.685396] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 2ee62ad3-0125-47dc-b163-7d15b6a17c8a] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 975.823978] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231152, 'name': Rename_Task, 'duration_secs': 0.237947} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.824545] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.824890] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52b8228a-0ca2-42d6-8296-8a40e394b6bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.834893] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 975.834893] env[70013]: value = "task-4231156" [ 975.834893] env[70013]: _type = "Task" [ 975.834893] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.844724] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.922987] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231155, 'name': CreateVM_Task, 'duration_secs': 0.385259} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.923186] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.923946] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.924132] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.924467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 975.924743] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b452b62a-69b8-412d-8d23-0512bcee4f1b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.933880] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 975.933880] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c08107-f541-8e3c-44fe-5a22723bc2d8" [ 975.933880] env[70013]: _type = "Task" [ 975.933880] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.951957] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c08107-f541-8e3c-44fe-5a22723bc2d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.952882] env[70013]: DEBUG nova.network.neutron [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Successfully updated port: 20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.965096] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17084b4-a339-416f-9490-7806cd32c2db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.973801] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff86e08-7e44-4007-b14a-b2f62667645a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.006616] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e602e800-9ced-42cf-96d5-8cf6ce7fba5d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.014687] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f477770-9bbc-4d6d-8f49-4ba1bf8a7259 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.029058] env[70013]: DEBUG nova.compute.provider_tree [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.192043] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: cfe1478c-53f2-4df8-99a8-ab23209f5306] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 976.200827] env[70013]: DEBUG nova.network.neutron [-] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.348105] env[70013]: DEBUG oslo_vmware.api [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231156, 'name': PowerOnVM_Task, 'duration_secs': 0.4649} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.348105] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.348105] env[70013]: INFO nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Took 14.85 seconds to spawn the instance on the hypervisor. [ 976.348325] env[70013]: DEBUG nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 976.351775] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8737f6-1789-4824-b8e8-cac5f6eaaf97 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.389710] env[70013]: DEBUG nova.network.neutron [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updated VIF entry in instance network info cache for port d8aacd8d-2fc2-477b-9d82-e101e20b0132. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 976.389938] env[70013]: DEBUG nova.network.neutron [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.444884] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c08107-f541-8e3c-44fe-5a22723bc2d8, 'name': SearchDatastore_Task, 'duration_secs': 0.01635} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.445210] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 976.445447] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.445683] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.445829] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 976.446019] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.446284] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b891f12a-87fc-47da-8160-6564a176577d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.455763] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.455763] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.456888] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abd4f2c3-00d1-4cda-8d40-f3b3d4b7835c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.459146] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.459317] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 976.459494] env[70013]: DEBUG nova.network.neutron [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.465832] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 976.465832] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a67447-e485-d2a4-24d3-932e60af3d44" [ 976.465832] env[70013]: _type = "Task" [ 976.465832] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.475165] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a67447-e485-d2a4-24d3-932e60af3d44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.532406] env[70013]: DEBUG nova.scheduler.client.report [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 976.695813] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 062eff58-95fe-4c9b-a586-7e7434c77adf] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 976.704359] env[70013]: INFO nova.compute.manager [-] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Took 1.36 seconds to deallocate network for instance. [ 976.870499] env[70013]: INFO nova.compute.manager [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Took 45.89 seconds to build instance. [ 976.894054] env[70013]: DEBUG oslo_concurrency.lockutils [req-9c319d5d-2fb3-4018-a29d-c80a06beca1c req-5a5881d1-2c39-408d-a8f3-9bbd8e138391 service nova] Releasing lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 976.978587] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a67447-e485-d2a4-24d3-932e60af3d44, 'name': SearchDatastore_Task, 'duration_secs': 0.01098} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.978893] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-034e5362-d567-42bf-bc90-b54fad64d2d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.985037] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 976.985037] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5204fcac-adb7-adff-3550-f677c9f51cfa" [ 976.985037] env[70013]: _type = "Task" [ 976.985037] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.994228] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5204fcac-adb7-adff-3550-f677c9f51cfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.995298] env[70013]: DEBUG nova.network.neutron [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.039350] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.043182] env[70013]: DEBUG nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 977.043182] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.685s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.044796] env[70013]: INFO nova.compute.claims [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.153166] env[70013]: DEBUG nova.network.neutron [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Updating instance_info_cache with network_info: [{"id": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "address": "fa:16:3e:1b:aa:bf", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20f2f376-c1", "ovs_interfaceid": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.198976] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 9a0ecd6c-5853-446d-8f41-b2ee51e5259a] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 977.212731] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.375875] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9b30ee6-f7a3-4f6e-b086-f74337a2953c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.794s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.497258] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5204fcac-adb7-adff-3550-f677c9f51cfa, 'name': SearchDatastore_Task, 'duration_secs': 0.011703} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.497258] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 977.497349] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.497615] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46bb47e0-6f69-4bae-8b85-ba142674c1b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.505338] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 977.505338] env[70013]: value = "task-4231157" [ 977.505338] env[70013]: _type = "Task" [ 977.505338] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.514781] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231157, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.550209] env[70013]: DEBUG nova.compute.utils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 977.553299] env[70013]: DEBUG nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 977.553602] env[70013]: DEBUG nova.network.neutron [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 977.609124] env[70013]: DEBUG nova.policy [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2601a89c7a09474bb4da0b707d9987f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3c45f05e8334dcca93a5c6d4ef4c68b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 977.657199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 977.657592] env[70013]: DEBUG nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Instance network_info: |[{"id": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "address": "fa:16:3e:1b:aa:bf", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20f2f376-c1", "ovs_interfaceid": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 977.658162] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:aa:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20f2f376-c1c7-4171-bc55-1ecdb0cc5d14', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.666035] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Creating folder: Project (a3c45f05e8334dcca93a5c6d4ef4c68b). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 977.666391] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d30e5c1-6acd-4f27-8e6f-06f6e172449d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.678792] env[70013]: DEBUG nova.compute.manager [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Received event network-vif-deleted-7beab378-23cf-4f47-8004-b917e0a1e482 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 977.679051] env[70013]: DEBUG nova.compute.manager [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Received event network-vif-plugged-20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 977.679810] env[70013]: DEBUG oslo_concurrency.lockutils [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] Acquiring lock "54d39dfd-6429-4009-8d26-22c1fa46243e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.680152] env[70013]: DEBUG oslo_concurrency.lockutils [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.680401] env[70013]: DEBUG oslo_concurrency.lockutils [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.680625] env[70013]: DEBUG nova.compute.manager [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] No waiting events found dispatching network-vif-plugged-20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 977.680804] env[70013]: WARNING nova.compute.manager [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Received unexpected event network-vif-plugged-20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 for instance with vm_state building and task_state spawning. [ 977.680972] env[70013]: DEBUG nova.compute.manager [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Received event network-changed-20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 977.681146] env[70013]: DEBUG nova.compute.manager [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Refreshing instance network info cache due to event network-changed-20f2f376-c1c7-4171-bc55-1ecdb0cc5d14. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 977.681341] env[70013]: DEBUG oslo_concurrency.lockutils [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] Acquiring lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.681612] env[70013]: DEBUG oslo_concurrency.lockutils [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] Acquired lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 977.681703] env[70013]: DEBUG nova.network.neutron [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Refreshing network info cache for port 20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.685309] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Created folder: Project (a3c45f05e8334dcca93a5c6d4ef4c68b) in parent group-v836999. [ 977.685309] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Creating folder: Instances. Parent ref: group-v837166. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 977.688030] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "8dbef727-13e9-4231-95a8-65d015ee13be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.688030] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.688030] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.688030] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.688030] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.688925] env[70013]: INFO nova.compute.manager [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Terminating instance [ 977.691654] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f79d7407-b140-414f-b581-944169bb9c58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.702224] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 3e938bb3-54a6-49f8-809d-d5aee0349eae] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 977.708400] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Created folder: Instances in parent group-v837166. [ 977.708611] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 977.708827] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.709414] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7ea06c8-0bf2-4679-9343-4ea0a9eb43ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.734063] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.734063] env[70013]: value = "task-4231160" [ 977.734063] env[70013]: _type = "Task" [ 977.734063] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.747362] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231160, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.880694] env[70013]: DEBUG nova.compute.manager [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 977.905093] env[70013]: DEBUG nova.network.neutron [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Successfully created port: 36c36a41-a7ab-45a3-a1c4-be24ecb013e7 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.016988] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231157, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.062460] env[70013]: DEBUG nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 978.195488] env[70013]: DEBUG nova.compute.manager [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 978.195742] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.196691] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337432ab-6723-458d-9025-4555d1f46912 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.207053] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: b23b4e67-67a6-4117-9a62-3c0dcd25e828] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 978.208260] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.208702] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64bdd29f-429b-4b70-850d-b27cb1c486af {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.217344] env[70013]: DEBUG oslo_vmware.api [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 978.217344] env[70013]: value = "task-4231161" [ 978.217344] env[70013]: _type = "Task" [ 978.217344] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.229735] env[70013]: DEBUG oslo_vmware.api [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.243156] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231160, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.404814] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.412724] env[70013]: DEBUG nova.network.neutron [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Updated VIF entry in instance network info cache for port 20f2f376-c1c7-4171-bc55-1ecdb0cc5d14. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.413103] env[70013]: DEBUG nova.network.neutron [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Updating instance_info_cache with network_info: [{"id": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "address": "fa:16:3e:1b:aa:bf", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20f2f376-c1", "ovs_interfaceid": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.517804] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231157, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588196} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.519317] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.519572] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.520427] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4183c2-645b-4fe3-97b9-ecbbb21f6acb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.524622] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82b8faae-e8f0-49b1-87f1-1582ba00385c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.531219] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2aaf45e-31ae-4430-9573-f67a88d550e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.536107] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 978.536107] env[70013]: value = "task-4231162" [ 978.536107] env[70013]: _type = "Task" [ 978.536107] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.578628] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f617d2a2-1e2a-4396-92d2-959d832e00ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.579280] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231162, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.585781] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b6f7cc-012c-467e-9cf9-68291a88e556 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.602845] env[70013]: DEBUG nova.compute.provider_tree [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.710950] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: b7cb13ef-25eb-4e83-9b53-d9f1ddc62b25] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 978.730077] env[70013]: DEBUG oslo_vmware.api [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231161, 'name': PowerOffVM_Task, 'duration_secs': 0.266793} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.730077] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.730077] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.730077] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e81b1f5-3274-4144-9524-90a6242b4463 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.742396] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231160, 'name': CreateVM_Task, 'duration_secs': 0.781019} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.742564] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 978.743265] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.743427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 978.743752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 978.743999] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3128f5bb-7b15-4e57-80a8-f062651b7022 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.749961] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 978.749961] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ea9388-3700-dcbd-71a4-f26ec095b0e7" [ 978.749961] env[70013]: _type = "Task" [ 978.749961] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.761021] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ea9388-3700-dcbd-71a4-f26ec095b0e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.799988] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.800208] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.800389] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleting the datastore file [datastore2] 8dbef727-13e9-4231-95a8-65d015ee13be {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.800700] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5ef0eaa-4be2-48da-a42e-0278c0a41020 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.807286] env[70013]: DEBUG oslo_vmware.api [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 978.807286] env[70013]: value = "task-4231164" [ 978.807286] env[70013]: _type = "Task" [ 978.807286] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.816366] env[70013]: DEBUG oslo_vmware.api [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.915612] env[70013]: DEBUG oslo_concurrency.lockutils [req-b78e3edb-1fd7-4d38-8048-d1265e8f5384 req-70f75aa0-ff8d-4a13-85da-a8cdca86e23f service nova] Releasing lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 979.046731] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231162, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077368} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.047054] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.047862] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f330045-991a-4662-ab33-d5c1fa2e28bd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.070225] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.070509] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36061fa6-177c-4228-80eb-4a912a19c93f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.086284] env[70013]: DEBUG nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 979.097301] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 979.097301] env[70013]: value = "task-4231165" [ 979.097301] env[70013]: _type = "Task" [ 979.097301] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.109225] env[70013]: DEBUG nova.scheduler.client.report [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.112720] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.122843] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 979.123126] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.123290] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 979.123474] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.123737] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 979.123817] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 979.123992] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 979.124198] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 979.124374] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 979.124540] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 979.125377] env[70013]: DEBUG nova.virt.hardware [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 979.126937] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf52be90-1814-4b36-9298-4e81d99be7f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.134605] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80cb1c7-6efe-4e58-8804-839899ec08f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.214443] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: fc0b7f87-8a36-440e-b158-30b9d9920572] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 979.261424] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ea9388-3700-dcbd-71a4-f26ec095b0e7, 'name': SearchDatastore_Task, 'duration_secs': 0.015751} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.261829] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 979.262347] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.262347] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.262482] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 979.262660] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.262932] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-364bf8bf-b66d-432b-802e-d28e10bfcae3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.273174] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.273379] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.274220] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2ff64d3-fad3-4819-a100-e10876b0c0eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.280248] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 979.280248] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52715f03-c9b8-b598-9e94-e1e22e729f30" [ 979.280248] env[70013]: _type = "Task" [ 979.280248] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.288728] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52715f03-c9b8-b598-9e94-e1e22e729f30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.318332] env[70013]: DEBUG oslo_vmware.api [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16664} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.318601] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.318780] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.318953] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.319141] env[70013]: INFO nova.compute.manager [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Took 1.12 seconds to destroy the instance on the hypervisor. [ 979.319380] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 979.319574] env[70013]: DEBUG nova.compute.manager [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 979.319671] env[70013]: DEBUG nova.network.neutron [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 979.506717] env[70013]: DEBUG nova.network.neutron [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Successfully updated port: 36c36a41-a7ab-45a3-a1c4-be24ecb013e7 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.607955] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231165, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.613961] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.614490] env[70013]: DEBUG nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 979.617244] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.663s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.617462] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.619695] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.802s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.623024] env[70013]: INFO nova.compute.claims [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.654214] env[70013]: INFO nova.scheduler.client.report [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Deleted allocations for instance 965480de-4fe2-48ce-94ad-0838d78239dd [ 979.717952] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 8ec4b86d-e83b-4b28-ac4b-d990177b0ba5] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 979.751369] env[70013]: DEBUG nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Received event network-vif-plugged-36c36a41-a7ab-45a3-a1c4-be24ecb013e7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 979.751369] env[70013]: DEBUG oslo_concurrency.lockutils [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] Acquiring lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.751369] env[70013]: DEBUG oslo_concurrency.lockutils [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.751369] env[70013]: DEBUG oslo_concurrency.lockutils [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.751369] env[70013]: DEBUG nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] No waiting events found dispatching network-vif-plugged-36c36a41-a7ab-45a3-a1c4-be24ecb013e7 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 979.751736] env[70013]: WARNING nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Received unexpected event network-vif-plugged-36c36a41-a7ab-45a3-a1c4-be24ecb013e7 for instance with vm_state building and task_state spawning. [ 979.751949] env[70013]: DEBUG nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Received event network-changed-36c36a41-a7ab-45a3-a1c4-be24ecb013e7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 979.752279] env[70013]: DEBUG nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Refreshing instance network info cache due to event network-changed-36c36a41-a7ab-45a3-a1c4-be24ecb013e7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 979.752498] env[70013]: DEBUG oslo_concurrency.lockutils [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] Acquiring lock "refresh_cache-c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.752641] env[70013]: DEBUG oslo_concurrency.lockutils [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] Acquired lock "refresh_cache-c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 979.752802] env[70013]: DEBUG nova.network.neutron [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Refreshing network info cache for port 36c36a41-a7ab-45a3-a1c4-be24ecb013e7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.790712] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52715f03-c9b8-b598-9e94-e1e22e729f30, 'name': SearchDatastore_Task, 'duration_secs': 0.018051} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.791623] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec33f5bd-9f0b-4553-96d4-1ef49b445263 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.797381] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 979.797381] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e7c16a-4ae4-5953-34e6-0d8ed3837c4f" [ 979.797381] env[70013]: _type = "Task" [ 979.797381] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.805665] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e7c16a-4ae4-5953-34e6-0d8ed3837c4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.010389] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "refresh_cache-c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.111128] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231165, 'name': ReconfigVM_Task, 'duration_secs': 0.534554} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.111128] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.111128] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78a59bc0-1668-4ec1-833d-eefa8dcbfdba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.118982] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 980.118982] env[70013]: value = "task-4231166" [ 980.118982] env[70013]: _type = "Task" [ 980.118982] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.128698] env[70013]: DEBUG nova.compute.utils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 980.131855] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231166, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.132450] env[70013]: DEBUG nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 980.132634] env[70013]: DEBUG nova.network.neutron [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 980.165626] env[70013]: DEBUG oslo_concurrency.lockutils [None req-08028261-2441-4592-8f5d-ef4d1d55ed4e tempest-ServersListShow2100Test-1069170016 tempest-ServersListShow2100Test-1069170016-project-member] Lock "965480de-4fe2-48ce-94ad-0838d78239dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.521s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 980.179452] env[70013]: DEBUG nova.network.neutron [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.197238] env[70013]: DEBUG nova.policy [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2601a89c7a09474bb4da0b707d9987f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3c45f05e8334dcca93a5c6d4ef4c68b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 980.221825] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 64308d7c-b63b-4fa6-b235-2f0586a888cf] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 980.310263] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e7c16a-4ae4-5953-34e6-0d8ed3837c4f, 'name': SearchDatastore_Task, 'duration_secs': 0.01049} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.313200] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 980.313200] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 54d39dfd-6429-4009-8d26-22c1fa46243e/54d39dfd-6429-4009-8d26-22c1fa46243e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.313200] env[70013]: DEBUG nova.network.neutron [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 980.317107] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84b2ad40-e8b8-4ace-b6cb-cec0b4933869 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.328021] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 980.328021] env[70013]: value = "task-4231167" [ 980.328021] env[70013]: _type = "Task" [ 980.328021] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.333719] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231167, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.453039] env[70013]: DEBUG nova.network.neutron [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.632485] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231166, 'name': Rename_Task, 'duration_secs': 0.165077} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.635914] env[70013]: DEBUG nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 980.638836] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 980.640449] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d6ae7d1-f2f4-44d4-a8b9-78337e1e8f8b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.646234] env[70013]: DEBUG nova.network.neutron [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Successfully created port: e7899a16-e277-4a87-8992-3900bdc7c327 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.649518] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 980.649518] env[70013]: value = "task-4231168" [ 980.649518] env[70013]: _type = "Task" [ 980.649518] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.660786] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.686504] env[70013]: INFO nova.compute.manager [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Took 1.37 seconds to deallocate network for instance. [ 980.727944] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 225c02d6-43df-4260-a428-14a6b1ed32cc] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 980.835754] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231167, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.958918] env[70013]: DEBUG oslo_concurrency.lockutils [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] Releasing lock "refresh_cache-c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 980.958918] env[70013]: DEBUG nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Received event network-vif-deleted-4edd213b-3ed8-4731-8128-4caf383b2277 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 980.958918] env[70013]: INFO nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Neutron deleted interface 4edd213b-3ed8-4731-8128-4caf383b2277; detaching it from the instance and deleting it from the info cache [ 980.958918] env[70013]: DEBUG nova.network.neutron [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.958918] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "refresh_cache-c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 980.958918] env[70013]: DEBUG nova.network.neutron [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.098849] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb60bd5b-d7d6-44b4-b2b4-eea60b29126e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.107406] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5dc7d5-fd03-4cea-ac5f-928e1d955326 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.144271] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb89e5c-2242-4a45-a4b3-2e88e9d34e2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.158991] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc199d0-8fc3-4867-b808-d5735f51238a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.167048] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231168, 'name': PowerOnVM_Task} progress is 90%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.178927] env[70013]: DEBUG nova.compute.provider_tree [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.194397] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.232175] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: cf8be667-da84-466b-9b7b-3bdfda6a8193] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 981.336786] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231167, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515457} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.338034] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 54d39dfd-6429-4009-8d26-22c1fa46243e/54d39dfd-6429-4009-8d26-22c1fa46243e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 981.338034] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.338034] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12c33c3c-14c1-4c96-8962-e336fa04be16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.345461] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 981.345461] env[70013]: value = "task-4231169" [ 981.345461] env[70013]: _type = "Task" [ 981.345461] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.354255] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231169, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.462181] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a5abcd1-0213-43af-afde-1ba44b5adf8d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.472226] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb14aa53-3df3-4fe6-9b98-c5f7e3872378 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.503494] env[70013]: DEBUG nova.compute.manager [req-9fc697d7-29b0-4af7-a7e9-8f6a203e93da req-056cd351-2e41-4a9d-b375-9c45d382f60a service nova] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Detach interface failed, port_id=4edd213b-3ed8-4731-8128-4caf383b2277, reason: Instance 8dbef727-13e9-4231-95a8-65d015ee13be could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 981.504601] env[70013]: DEBUG nova.network.neutron [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 981.649266] env[70013]: DEBUG nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 981.663669] env[70013]: DEBUG oslo_vmware.api [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231168, 'name': PowerOnVM_Task, 'duration_secs': 0.610443} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.663927] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.664189] env[70013]: INFO nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Took 8.92 seconds to spawn the instance on the hypervisor. [ 981.664407] env[70013]: DEBUG nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 981.665646] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bfaf60-d479-47e6-9b6b-d0e3b13c8165 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.682357] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 981.682623] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.682785] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 981.682969] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.683142] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 981.683270] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 981.683504] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 981.683679] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 981.683846] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 981.684014] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 981.684199] env[70013]: DEBUG nova.virt.hardware [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 981.685495] env[70013]: DEBUG nova.scheduler.client.report [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 981.692941] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53efd5f7-0b88-4596-9c7c-99157b67da03 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.701203] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ddf47d-2159-4811-87d1-963f46887a48 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.718246] env[70013]: DEBUG nova.network.neutron [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Updating instance_info_cache with network_info: [{"id": "36c36a41-a7ab-45a3-a1c4-be24ecb013e7", "address": "fa:16:3e:d8:b1:52", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36c36a41-a7", "ovs_interfaceid": "36c36a41-a7ab-45a3-a1c4-be24ecb013e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.736642] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 9a553d85-5086-46c5-8df4-451928e38ed9] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 981.859025] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231169, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075694} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.859025] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 981.859025] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d006a219-a762-43f2-ba0b-cea8a87cd0c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.880976] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 54d39dfd-6429-4009-8d26-22c1fa46243e/54d39dfd-6429-4009-8d26-22c1fa46243e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.881301] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a19f755-90d2-4271-85a7-7a81148abc62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.902090] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 981.902090] env[70013]: value = "task-4231170" [ 981.902090] env[70013]: _type = "Task" [ 981.902090] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.910494] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231170, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.201931] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.582s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.205352] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.145s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.206962] env[70013]: INFO nova.compute.claims [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.213907] env[70013]: INFO nova.compute.manager [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Took 43.91 seconds to build instance. [ 982.213907] env[70013]: DEBUG nova.network.neutron [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Successfully updated port: e7899a16-e277-4a87-8992-3900bdc7c327 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.224023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "refresh_cache-c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 982.224023] env[70013]: DEBUG nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Instance network_info: |[{"id": "36c36a41-a7ab-45a3-a1c4-be24ecb013e7", "address": "fa:16:3e:d8:b1:52", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36c36a41-a7", "ovs_interfaceid": "36c36a41-a7ab-45a3-a1c4-be24ecb013e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 982.224023] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:b1:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '36c36a41-a7ab-45a3-a1c4-be24ecb013e7', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.231199] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 982.232134] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.232635] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80267261-8eea-4975-a081-db3ef003250b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.258873] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 3b29c99a-0d55-40b1-a155-199b1ced2146] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 982.269734] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.269734] env[70013]: value = "task-4231171" [ 982.269734] env[70013]: _type = "Task" [ 982.269734] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.284037] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231171, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.339168] env[70013]: DEBUG nova.compute.manager [req-1d76b177-0bcc-4b02-b052-e48e7f6d9465 req-2fb9a3f6-c5fc-4142-b5e4-1c037b0bd479 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Received event network-vif-plugged-e7899a16-e277-4a87-8992-3900bdc7c327 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 982.340733] env[70013]: DEBUG oslo_concurrency.lockutils [req-1d76b177-0bcc-4b02-b052-e48e7f6d9465 req-2fb9a3f6-c5fc-4142-b5e4-1c037b0bd479 service nova] Acquiring lock "9a61dedd-3764-4bd9-a300-480cc7d14a21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.341009] env[70013]: DEBUG oslo_concurrency.lockutils [req-1d76b177-0bcc-4b02-b052-e48e7f6d9465 req-2fb9a3f6-c5fc-4142-b5e4-1c037b0bd479 service nova] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.341198] env[70013]: DEBUG oslo_concurrency.lockutils [req-1d76b177-0bcc-4b02-b052-e48e7f6d9465 req-2fb9a3f6-c5fc-4142-b5e4-1c037b0bd479 service nova] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.341390] env[70013]: DEBUG nova.compute.manager [req-1d76b177-0bcc-4b02-b052-e48e7f6d9465 req-2fb9a3f6-c5fc-4142-b5e4-1c037b0bd479 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] No waiting events found dispatching network-vif-plugged-e7899a16-e277-4a87-8992-3900bdc7c327 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 982.341615] env[70013]: WARNING nova.compute.manager [req-1d76b177-0bcc-4b02-b052-e48e7f6d9465 req-2fb9a3f6-c5fc-4142-b5e4-1c037b0bd479 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Received unexpected event network-vif-plugged-e7899a16-e277-4a87-8992-3900bdc7c327 for instance with vm_state building and task_state spawning. [ 982.412418] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231170, 'name': ReconfigVM_Task, 'duration_secs': 0.345277} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.412767] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 54d39dfd-6429-4009-8d26-22c1fa46243e/54d39dfd-6429-4009-8d26-22c1fa46243e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.413398] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00d37a3f-8c7b-4041-9f53-a756433e328c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.422025] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 982.422025] env[70013]: value = "task-4231172" [ 982.422025] env[70013]: _type = "Task" [ 982.422025] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.435411] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231172, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.715845] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "7c199ad6-9d20-4c6a-ac96-21bc6e500155" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.716162] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "7c199ad6-9d20-4c6a-ac96-21bc6e500155" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.722989] env[70013]: DEBUG oslo_concurrency.lockutils [None req-773c58a4-e339-4677-b528-057e2228aea6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.531s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.723613] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "refresh_cache-9a61dedd-3764-4bd9-a300-480cc7d14a21" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.723753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "refresh_cache-9a61dedd-3764-4bd9-a300-480cc7d14a21" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 982.724188] env[70013]: DEBUG nova.network.neutron [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.725391] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "7c199ad6-9d20-4c6a-ac96-21bc6e500155" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.009s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.725967] env[70013]: DEBUG nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 982.763213] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 417f8c74-81cc-47d6-80ca-b61d03ebec50] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 982.782862] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231171, 'name': CreateVM_Task, 'duration_secs': 0.362603} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.783106] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.784195] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.784281] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 982.785028] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 982.785327] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a0cbe7e-b3d8-4ad9-b71d-70cc08bb861e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.792558] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 982.792558] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526d41c5-4eff-a995-8b29-4db38f59345a" [ 982.792558] env[70013]: _type = "Task" [ 982.792558] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.805982] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526d41c5-4eff-a995-8b29-4db38f59345a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.932776] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231172, 'name': Rename_Task, 'duration_secs': 0.170029} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.933357] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.933722] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cf875da-04f0-4f1e-9eff-ee1b577d00e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.940608] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 982.940608] env[70013]: value = "task-4231173" [ 982.940608] env[70013]: _type = "Task" [ 982.940608] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.949420] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231173, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.171098] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.171098] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.171098] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.171690] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.172446] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.174682] env[70013]: INFO nova.compute.manager [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Terminating instance [ 983.233097] env[70013]: DEBUG nova.compute.manager [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 983.238720] env[70013]: DEBUG nova.compute.utils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 983.242078] env[70013]: DEBUG nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 983.242078] env[70013]: DEBUG nova.network.neutron [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.269902] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 9c5f2422-820e-454b-98be-695e3ac6d1e3] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 983.299625] env[70013]: DEBUG nova.network.neutron [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 983.313472] env[70013]: DEBUG nova.policy [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c887d102aaab43abbfac897905570f82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bddca0c555f41d5ae6680737dc25fb6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 983.320144] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526d41c5-4eff-a995-8b29-4db38f59345a, 'name': SearchDatastore_Task, 'duration_secs': 0.017743} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.320753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 983.320999] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.321270] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.321420] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 983.322549] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.322549] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53ac1111-6737-4aa9-8de2-a68f2e329788 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.332234] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.332319] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.333224] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1008df5-cdc3-476b-914e-91e17b0a3a0e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.346088] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 983.346088] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522c91e8-6ae9-676e-058d-21191b21971d" [ 983.346088] env[70013]: _type = "Task" [ 983.346088] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.360624] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522c91e8-6ae9-676e-058d-21191b21971d, 'name': SearchDatastore_Task, 'duration_secs': 0.013117} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.361217] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3517089c-c7ba-4092-b741-509504f7fd1f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.370886] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 983.370886] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1829e-a938-6168-dfe2-cc00bb9f5bf8" [ 983.370886] env[70013]: _type = "Task" [ 983.370886] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.377574] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1829e-a938-6168-dfe2-cc00bb9f5bf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.453057] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231173, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.583266] env[70013]: DEBUG nova.network.neutron [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Updating instance_info_cache with network_info: [{"id": "e7899a16-e277-4a87-8992-3900bdc7c327", "address": "fa:16:3e:03:3d:bc", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7899a16-e2", "ovs_interfaceid": "e7899a16-e277-4a87-8992-3900bdc7c327", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.681040] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.681040] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquired lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 983.681040] env[70013]: DEBUG nova.network.neutron [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 983.706493] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.707438] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.735195] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10b3b17-0107-41e4-b512-26db02c1815b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.744662] env[70013]: DEBUG nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 983.749212] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b36dbe-372d-47d7-90c5-29e7c835eb53 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.816793] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: e8998cab-533b-46b1-a1ea-5eb5db7d6d99] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 983.821711] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.824411] env[70013]: DEBUG nova.network.neutron [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Successfully created port: ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.828759] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d074c15c-6158-4b84-93e2-f1d6a573fe5c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.842514] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "1230ca35-1512-4464-85fc-d3b4ab05eac1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.842952] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.852193] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb384761-d30d-423c-8756-54db5146b29f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.873488] env[70013]: DEBUG nova.compute.provider_tree [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.885300] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d1829e-a938-6168-dfe2-cc00bb9f5bf8, 'name': SearchDatastore_Task, 'duration_secs': 0.019239} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.886218] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 983.886486] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3/c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 983.886764] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-026b4b27-6b44-40cb-b974-cdf3b49a4f89 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.895903] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 983.895903] env[70013]: value = "task-4231174" [ 983.895903] env[70013]: _type = "Task" [ 983.895903] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.908660] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231174, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.953799] env[70013]: DEBUG oslo_vmware.api [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231173, 'name': PowerOnVM_Task, 'duration_secs': 0.85071} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.954035] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.954232] env[70013]: INFO nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Took 8.48 seconds to spawn the instance on the hypervisor. [ 983.954438] env[70013]: DEBUG nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 983.955253] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de56d3f5-68cf-4387-b5d5-7c648b076066 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.086022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "refresh_cache-9a61dedd-3764-4bd9-a300-480cc7d14a21" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 984.086473] env[70013]: DEBUG nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Instance network_info: |[{"id": "e7899a16-e277-4a87-8992-3900bdc7c327", "address": "fa:16:3e:03:3d:bc", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7899a16-e2", "ovs_interfaceid": "e7899a16-e277-4a87-8992-3900bdc7c327", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 984.087296] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:3d:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7899a16-e277-4a87-8992-3900bdc7c327', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.095770] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 984.096061] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.096675] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ec6c205a-d210-4ce3-a6d3-06765813e2bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.120103] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.120103] env[70013]: value = "task-4231175" [ 984.120103] env[70013]: _type = "Task" [ 984.120103] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.129652] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231175, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.203941] env[70013]: DEBUG nova.network.neutron [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.263490] env[70013]: DEBUG nova.network.neutron [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.323214] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 98a0199a-4209-4faf-adf5-7ae33b099d20] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 984.367417] env[70013]: DEBUG nova.compute.manager [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Received event network-changed-e7899a16-e277-4a87-8992-3900bdc7c327 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 984.367509] env[70013]: DEBUG nova.compute.manager [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Refreshing instance network info cache due to event network-changed-e7899a16-e277-4a87-8992-3900bdc7c327. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 984.367928] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] Acquiring lock "refresh_cache-9a61dedd-3764-4bd9-a300-480cc7d14a21" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.367928] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] Acquired lock "refresh_cache-9a61dedd-3764-4bd9-a300-480cc7d14a21" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 984.368117] env[70013]: DEBUG nova.network.neutron [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Refreshing network info cache for port e7899a16-e277-4a87-8992-3900bdc7c327 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 984.379756] env[70013]: DEBUG nova.scheduler.client.report [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 984.412224] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231174, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.474528] env[70013]: INFO nova.compute.manager [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Took 45.13 seconds to build instance. [ 984.630559] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231175, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.763400] env[70013]: DEBUG nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 984.769616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Releasing lock "refresh_cache-85488f5e-bd8a-4bcf-8420-744c54176c2e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 984.770072] env[70013]: DEBUG nova.compute.manager [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 984.770264] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.770580] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a450fe9-b364-4b0a-ae0c-42cd4cb5e2ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.786924] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98189256-6721-4e58-9867-f303feb593f9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.807934] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 984.807934] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.808279] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 984.808279] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.808388] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 984.809060] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 984.809060] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 984.809060] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 984.809257] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 984.809348] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 984.809524] env[70013]: DEBUG nova.virt.hardware [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 984.810659] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e50ebc5-a4ec-44ad-8cb2-9fdfce0991d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.824994] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 85488f5e-bd8a-4bcf-8420-744c54176c2e could not be found. [ 984.825312] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 984.825409] env[70013]: INFO nova.compute.manager [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Took 0.06 seconds to destroy the instance on the hypervisor. [ 984.825662] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 984.826501] env[70013]: DEBUG nova.compute.manager [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 984.826615] env[70013]: DEBUG nova.network.neutron [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.828684] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 2b91ec52-2d32-4f86-acf2-f74f661f65b5] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 984.837151] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2e5a0a-2f10-47f2-8b6d-452c7da24e5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.858980] env[70013]: DEBUG nova.network.neutron [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.885282] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.680s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.885889] env[70013]: DEBUG nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 984.892120] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.181s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.908331] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231174, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629551} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.908331] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3/c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 984.908331] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.908331] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7affcd4a-52ca-4999-ad1b-97c0a257f4e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.921023] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 984.921023] env[70013]: value = "task-4231176" [ 984.921023] env[70013]: _type = "Task" [ 984.921023] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.935352] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231176, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.978645] env[70013]: DEBUG oslo_concurrency.lockutils [None req-232e62ce-fba2-40b7-b09e-9c8d1db96fda tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.192s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 985.135066] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231175, 'name': CreateVM_Task, 'duration_secs': 0.834902} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.135833] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.136540] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.136710] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.137057] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 985.137310] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c6aae4b-2921-4959-ab10-51908f609050 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.144370] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 985.144370] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52964cd5-10f7-a990-ed5a-95304007ff01" [ 985.144370] env[70013]: _type = "Task" [ 985.144370] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.152555] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52964cd5-10f7-a990-ed5a-95304007ff01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.220159] env[70013]: DEBUG nova.network.neutron [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Updated VIF entry in instance network info cache for port e7899a16-e277-4a87-8992-3900bdc7c327. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 985.220549] env[70013]: DEBUG nova.network.neutron [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Updating instance_info_cache with network_info: [{"id": "e7899a16-e277-4a87-8992-3900bdc7c327", "address": "fa:16:3e:03:3d:bc", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7899a16-e2", "ovs_interfaceid": "e7899a16-e277-4a87-8992-3900bdc7c327", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.334856] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.334973] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Cleaning up deleted instances with incomplete migration {{(pid=70013) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11902}} [ 985.363999] env[70013]: DEBUG nova.network.neutron [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.368209] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e94a07-25ac-4519-8893-cd17b9632086 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.376943] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28537245-856a-4661-b656-74631b2be43d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.409563] env[70013]: DEBUG nova.compute.utils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 985.412143] env[70013]: DEBUG nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 985.412143] env[70013]: DEBUG nova.network.neutron [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.417984] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bbfc77-0d61-4ace-890d-d42bbd901798 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.417984] env[70013]: DEBUG nova.network.neutron [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Successfully updated port: ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.428834] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fe2fec-f35d-4ea2-90eb-972c88ecb328 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.438690] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079714} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.439344] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.440121] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b01b6f-76e2-46f2-9ffe-69028b148980 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.451070] env[70013]: DEBUG nova.compute.provider_tree [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.474945] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3/c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.476629] env[70013]: DEBUG nova.policy [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfa9f992483942f689f616b60a27b00d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0896b76438d9456fb1b9fa2f39d0decc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 985.478406] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35e55c78-03db-4575-a05e-ffc7864feb07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.493129] env[70013]: DEBUG nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 985.504144] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 985.504144] env[70013]: value = "task-4231177" [ 985.504144] env[70013]: _type = "Task" [ 985.504144] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.518010] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231177, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.656420] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52964cd5-10f7-a990-ed5a-95304007ff01, 'name': SearchDatastore_Task, 'duration_secs': 0.052777} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.656776] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 985.657066] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.657333] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.657494] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.657684] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.657973] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d70a8ad6-e481-4278-b29b-0b56733866ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.669211] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.669395] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.670203] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17dc1df5-477d-44d6-9947-8d08b20e0927 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.677115] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 985.677115] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]524acc46-57c3-04eb-2359-66b02ca7cb32" [ 985.677115] env[70013]: _type = "Task" [ 985.677115] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.685834] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524acc46-57c3-04eb-2359-66b02ca7cb32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.723856] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f9ef324-3af0-4bdd-a9f7-6aa28845ae9e req-de6159c1-394f-4a4c-b1b8-35f808d72594 service nova] Releasing lock "refresh_cache-9a61dedd-3764-4bd9-a300-480cc7d14a21" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 985.801429] env[70013]: DEBUG nova.network.neutron [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Successfully created port: a9940c14-e6da-4fa5-b093-ba598b0bdc93 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.837942] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.867501] env[70013]: INFO nova.compute.manager [-] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Took 1.04 seconds to deallocate network for instance. [ 985.913511] env[70013]: DEBUG nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 985.921962] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "refresh_cache-b292a985-3d2a-4cf3-a09b-8e72d8d21078" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.921962] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquired lock "refresh_cache-b292a985-3d2a-4cf3-a09b-8e72d8d21078" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.921962] env[70013]: DEBUG nova.network.neutron [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.955192] env[70013]: DEBUG nova.scheduler.client.report [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 985.956390] env[70013]: DEBUG oslo_concurrency.lockutils [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.067s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 985.956600] env[70013]: INFO nova.compute.manager [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 2aefaa91-c439-486a-8b19-c6f45f52583f] Successfully reverted task state from None on failure for instance. [ 985.959460] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.184s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 985.959645] env[70013]: DEBUG nova.objects.instance [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server [None req-508b6676-f9ca-49a9-8421-57db8e3c63d8 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server yield [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6cf2dd0b-e8f3-4eb5-b1b8-77cbbb05a812"}]} [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 985.963572] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 985.964849] env[70013]: ERROR oslo_messaging.rpc.server [ 986.020174] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231177, 'name': ReconfigVM_Task, 'duration_secs': 0.335264} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.020174] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Reconfigured VM instance instance-00000048 to attach disk [datastore2] c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3/c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.020174] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 986.020174] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53e7c945-a1a1-42f5-bc04-26c16c38f3b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.028078] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 986.028078] env[70013]: value = "task-4231178" [ 986.028078] env[70013]: _type = "Task" [ 986.028078] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.038020] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231178, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.192883] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524acc46-57c3-04eb-2359-66b02ca7cb32, 'name': SearchDatastore_Task, 'duration_secs': 0.018401} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.193818] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6259aad4-ef38-407c-887c-e6df47157eaf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.200339] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 986.200339] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e19192-6f14-973e-e4fa-823c006d397f" [ 986.200339] env[70013]: _type = "Task" [ 986.200339] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.210029] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e19192-6f14-973e-e4fa-823c006d397f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.375744] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 986.415394] env[70013]: DEBUG nova.compute.manager [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Received event network-vif-plugged-ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 986.415622] env[70013]: DEBUG oslo_concurrency.lockutils [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] Acquiring lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 986.415837] env[70013]: DEBUG oslo_concurrency.lockutils [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 986.416008] env[70013]: DEBUG oslo_concurrency.lockutils [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 986.416212] env[70013]: DEBUG nova.compute.manager [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] No waiting events found dispatching network-vif-plugged-ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 986.416431] env[70013]: WARNING nova.compute.manager [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Received unexpected event network-vif-plugged-ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 for instance with vm_state building and task_state spawning. [ 986.416601] env[70013]: DEBUG nova.compute.manager [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Received event network-changed-ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 986.416754] env[70013]: DEBUG nova.compute.manager [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Refreshing instance network info cache due to event network-changed-ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 986.416921] env[70013]: DEBUG oslo_concurrency.lockutils [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] Acquiring lock "refresh_cache-b292a985-3d2a-4cf3-a09b-8e72d8d21078" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.458037] env[70013]: DEBUG nova.network.neutron [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 986.540261] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231178, 'name': Rename_Task, 'duration_secs': 0.165508} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.540653] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 986.540916] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2227c3ed-c5d6-48c0-bf7d-d3bedf593086 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.552765] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 986.552765] env[70013]: value = "task-4231179" [ 986.552765] env[70013]: _type = "Task" [ 986.552765] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.563178] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231179, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.622759] env[70013]: DEBUG nova.network.neutron [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Updating instance_info_cache with network_info: [{"id": "ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7", "address": "fa:16:3e:4f:8f:4e", "network": {"id": "f6d5a925-226e-4e70-881e-3623771aff58", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-990756005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bddca0c555f41d5ae6680737dc25fb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "310b8ba9-edca-4135-863e-f4a786dd4a77", "external-id": "nsx-vlan-transportzone-768", "segmentation_id": 768, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccf0fb0e-eb", "ovs_interfaceid": "ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.712962] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e19192-6f14-973e-e4fa-823c006d397f, 'name': SearchDatastore_Task, 'duration_secs': 0.017378} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.713302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 986.713578] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 9a61dedd-3764-4bd9-a300-480cc7d14a21/9a61dedd-3764-4bd9-a300-480cc7d14a21.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.713890] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7faa398-90a8-4cb7-b1ea-bd869a6ef124 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.724109] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 986.724109] env[70013]: value = "task-4231180" [ 986.724109] env[70013]: _type = "Task" [ 986.724109] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.734143] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.927253] env[70013]: DEBUG nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 986.953073] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 986.953337] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.953551] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 986.953821] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.954063] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 986.954387] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 986.954562] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 986.954782] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 986.955066] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 986.955238] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 986.955444] env[70013]: DEBUG nova.virt.hardware [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 986.956537] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279f7cca-252e-47a5-9028-942f5c8cc8f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.968305] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813fc070-139f-4f45-8d98-49fcfb4a3eef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.975030] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fa621ba3-d807-4da4-9bea-ebc57cc7628d tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 986.976192] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.760s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 986.977996] env[70013]: INFO nova.compute.claims [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.067841] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231179, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.129169] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Releasing lock "refresh_cache-b292a985-3d2a-4cf3-a09b-8e72d8d21078" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 987.129169] env[70013]: DEBUG nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Instance network_info: |[{"id": "ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7", "address": "fa:16:3e:4f:8f:4e", "network": {"id": "f6d5a925-226e-4e70-881e-3623771aff58", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-990756005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bddca0c555f41d5ae6680737dc25fb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "310b8ba9-edca-4135-863e-f4a786dd4a77", "external-id": "nsx-vlan-transportzone-768", "segmentation_id": 768, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccf0fb0e-eb", "ovs_interfaceid": "ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 987.129169] env[70013]: DEBUG oslo_concurrency.lockutils [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] Acquired lock "refresh_cache-b292a985-3d2a-4cf3-a09b-8e72d8d21078" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 987.129169] env[70013]: DEBUG nova.network.neutron [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Refreshing network info cache for port ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.129169] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:8f:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '310b8ba9-edca-4135-863e-f4a786dd4a77', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.143314] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Creating folder: Project (6bddca0c555f41d5ae6680737dc25fb6). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.148461] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27431525-7655-4d9d-a69c-7af2167a3779 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.164697] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Created folder: Project (6bddca0c555f41d5ae6680737dc25fb6) in parent group-v836999. [ 987.165472] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Creating folder: Instances. Parent ref: group-v837171. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.165472] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4dc1c37-f983-4da6-8c70-c82730433bc0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.182654] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Created folder: Instances in parent group-v837171. [ 987.183118] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 987.183544] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.183641] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d185791-86d5-447c-b48b-c0cb625440d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.212192] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.212192] env[70013]: value = "task-4231183" [ 987.212192] env[70013]: _type = "Task" [ 987.212192] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.232678] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231183, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.239745] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231180, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.536794] env[70013]: DEBUG nova.network.neutron [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Updated VIF entry in instance network info cache for port ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.537265] env[70013]: DEBUG nova.network.neutron [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Updating instance_info_cache with network_info: [{"id": "ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7", "address": "fa:16:3e:4f:8f:4e", "network": {"id": "f6d5a925-226e-4e70-881e-3623771aff58", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-990756005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bddca0c555f41d5ae6680737dc25fb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "310b8ba9-edca-4135-863e-f4a786dd4a77", "external-id": "nsx-vlan-transportzone-768", "segmentation_id": 768, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccf0fb0e-eb", "ovs_interfaceid": "ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.566311] env[70013]: DEBUG oslo_vmware.api [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231179, 'name': PowerOnVM_Task, 'duration_secs': 0.646379} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.568371] env[70013]: DEBUG nova.network.neutron [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Successfully updated port: a9940c14-e6da-4fa5-b093-ba598b0bdc93 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.569512] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 987.569773] env[70013]: INFO nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Took 8.48 seconds to spawn the instance on the hypervisor. [ 987.569998] env[70013]: DEBUG nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 987.571260] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3ae94f-6e26-4dcc-8720-6803d304fad5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.721829] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231183, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.736041] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231180, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616146} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.736041] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 9a61dedd-3764-4bd9-a300-480cc7d14a21/9a61dedd-3764-4bd9-a300-480cc7d14a21.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 987.736041] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.736394] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-013e90ef-d40f-4f32-9984-f43cae73bdc2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.745256] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 987.745256] env[70013]: value = "task-4231184" [ 987.745256] env[70013]: _type = "Task" [ 987.745256] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.756015] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.040104] env[70013]: DEBUG oslo_concurrency.lockutils [req-aa7bb96a-d176-4e3f-bb23-e3158c62eb8a req-6e8e78de-3d0d-42a4-a049-428c2e00e08b service nova] Releasing lock "refresh_cache-b292a985-3d2a-4cf3-a09b-8e72d8d21078" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.071705] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-2070989d-6c90-4eb1-8508-7587ed9659d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.072203] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-2070989d-6c90-4eb1-8508-7587ed9659d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.072377] env[70013]: DEBUG nova.network.neutron [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.097928] env[70013]: INFO nova.compute.manager [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Took 43.76 seconds to build instance. [ 988.225727] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231183, 'name': CreateVM_Task, 'duration_secs': 0.65255} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.225929] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.226996] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.227206] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.227658] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 988.227922] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6e3d667-44b1-4d2c-8d2e-55d4cb637e3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.241193] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 988.241193] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522192df-a975-b1ca-205b-7ded9f4e3a9b" [ 988.241193] env[70013]: _type = "Task" [ 988.241193] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.256122] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522192df-a975-b1ca-205b-7ded9f4e3a9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.262413] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.183365} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.262881] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.263978] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1923207a-b3ad-4377-9676-f415004441d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.293850] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 9a61dedd-3764-4bd9-a300-480cc7d14a21/9a61dedd-3764-4bd9-a300-480cc7d14a21.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.296768] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d52594c-44c6-4d74-b15a-9d8024b711f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.318964] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 988.318964] env[70013]: value = "task-4231185" [ 988.318964] env[70013]: _type = "Task" [ 988.318964] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.330231] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231185, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.444125] env[70013]: DEBUG nova.compute.manager [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Received event network-vif-plugged-a9940c14-e6da-4fa5-b093-ba598b0bdc93 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 988.444346] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] Acquiring lock "2070989d-6c90-4eb1-8508-7587ed9659d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.444558] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.444722] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 988.444947] env[70013]: DEBUG nova.compute.manager [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] No waiting events found dispatching network-vif-plugged-a9940c14-e6da-4fa5-b093-ba598b0bdc93 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 988.445115] env[70013]: WARNING nova.compute.manager [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Received unexpected event network-vif-plugged-a9940c14-e6da-4fa5-b093-ba598b0bdc93 for instance with vm_state building and task_state spawning. [ 988.445207] env[70013]: DEBUG nova.compute.manager [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Received event network-changed-a9940c14-e6da-4fa5-b093-ba598b0bdc93 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 988.445356] env[70013]: DEBUG nova.compute.manager [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Refreshing instance network info cache due to event network-changed-a9940c14-e6da-4fa5-b093-ba598b0bdc93. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 988.445557] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] Acquiring lock "refresh_cache-2070989d-6c90-4eb1-8508-7587ed9659d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.460405] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56dffe9-3c52-4459-82ba-1e80d8728fa3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.468933] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6e9883-0f8d-4816-be6f-b8ee3b097cc9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.500120] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0b9b30-40e9-4e16-b154-bd80a80140c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.508195] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f945efa8-3067-4f74-a413-8d2effff2cad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.524635] env[70013]: DEBUG nova.compute.provider_tree [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.600192] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a95340f5-5734-4e1e-a18c-04844bbe1b85 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.422s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 988.635730] env[70013]: DEBUG nova.network.neutron [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.753948] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522192df-a975-b1ca-205b-7ded9f4e3a9b, 'name': SearchDatastore_Task, 'duration_secs': 0.063469} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.754781] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.755159] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.755510] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.755774] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.756101] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.756455] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5cc1ac34-a3da-4fd2-a61b-06c16ce38f5d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.770119] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.770119] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.770326] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-923a2054-3c41-4f8e-8b46-3c006b563ecc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.778331] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 988.778331] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522a25b4-db54-a98f-c1b9-085363c6bb66" [ 988.778331] env[70013]: _type = "Task" [ 988.778331] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.786213] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522a25b4-db54-a98f-c1b9-085363c6bb66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.829291] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231185, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.017501] env[70013]: DEBUG nova.network.neutron [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Updating instance_info_cache with network_info: [{"id": "a9940c14-e6da-4fa5-b093-ba598b0bdc93", "address": "fa:16:3e:08:d8:08", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9940c14-e6", "ovs_interfaceid": "a9940c14-e6da-4fa5-b093-ba598b0bdc93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.054255] env[70013]: ERROR nova.scheduler.client.report [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [req-f1160601-5395-4761-83ad-a4facb0279a4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f1160601-5395-4761-83ad-a4facb0279a4"}]} [ 989.054733] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 989.055807] env[70013]: ERROR nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Traceback (most recent call last): [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] yield [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] self.set_inventory_for_provider( [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f1160601-5395-4761-83ad-a4facb0279a4"}]} [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] During handling of the above exception, another exception occurred: [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Traceback (most recent call last): [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] with self.rt.instance_claim(context, instance, node, allocs, [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] return f(*args, **kwargs) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] self._update(elevated, cn) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] self._update_to_placement(context, compute_node, startup) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] return attempt.get(self._wrap_exception) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] six.reraise(self.value[0], self.value[1], self.value[2]) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] raise value [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] self.reportclient.update_from_provider_tree( [ 989.055807] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 989.056930] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] with catch_all(pd.uuid): [ 989.056930] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 989.056930] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] self.gen.throw(typ, value, traceback) [ 989.056930] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 989.056930] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] raise exception.ResourceProviderSyncFailed() [ 989.056930] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 989.056930] env[70013]: ERROR nova.compute.manager [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] [ 989.056930] env[70013]: DEBUG nova.compute.utils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 989.058782] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.358s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 989.062064] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Build of instance 69f67e8d-2da0-474c-92db-8388c98b61f4 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 989.062479] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 989.062719] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "refresh_cache-69f67e8d-2da0-474c-92db-8388c98b61f4" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.062868] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "refresh_cache-69f67e8d-2da0-474c-92db-8388c98b61f4" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 989.063283] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.092764] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 989.104875] env[70013]: DEBUG nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 989.109503] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 989.109751] env[70013]: DEBUG nova.compute.provider_tree [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.123564] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 989.146515] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 989.288849] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522a25b4-db54-a98f-c1b9-085363c6bb66, 'name': SearchDatastore_Task, 'duration_secs': 0.040635} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.292457] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87a8e8bd-0e06-4ff4-9a5c-0ce74b943621 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.301046] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 989.301046] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52105e4e-2375-243f-fafa-ff7d996e52e1" [ 989.301046] env[70013]: _type = "Task" [ 989.301046] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.314938] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52105e4e-2375-243f-fafa-ff7d996e52e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.333540] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231185, 'name': ReconfigVM_Task, 'duration_secs': 0.849625} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.333745] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 9a61dedd-3764-4bd9-a300-480cc7d14a21/9a61dedd-3764-4bd9-a300-480cc7d14a21.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.334472] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07d8dc8b-4f2c-4ac6-aeec-54e3c854309b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.345370] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 989.345370] env[70013]: value = "task-4231186" [ 989.345370] env[70013]: _type = "Task" [ 989.345370] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.355127] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231186, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.520955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-2070989d-6c90-4eb1-8508-7587ed9659d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 989.521437] env[70013]: DEBUG nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Instance network_info: |[{"id": "a9940c14-e6da-4fa5-b093-ba598b0bdc93", "address": "fa:16:3e:08:d8:08", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9940c14-e6", "ovs_interfaceid": "a9940c14-e6da-4fa5-b093-ba598b0bdc93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 989.522746] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] Acquired lock "refresh_cache-2070989d-6c90-4eb1-8508-7587ed9659d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 989.523100] env[70013]: DEBUG nova.network.neutron [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Refreshing network info cache for port a9940c14-e6da-4fa5-b093-ba598b0bdc93 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.525049] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:d8:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac7039c0-3374-4c08-87fc-af2449b48b02', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a9940c14-e6da-4fa5-b093-ba598b0bdc93', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.533713] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 989.539936] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.540685] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9ae175f-9eda-48ea-9ba9-f4402eb65344 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.563762] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.563762] env[70013]: value = "task-4231187" [ 989.563762] env[70013]: _type = "Task" [ 989.563762] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.578476] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231187, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.598371] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.636840] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.713758] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7058c9-ddf0-42d3-a7f5-792cc2558f91 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.723427] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefccc54-5ea9-4c0b-b0b5-7828a131545e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.761900] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38d1242-3760-4904-a297-d8c4de36105d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.771572] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc67e9c-63a1-44c7-b034-1a522c788a1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.787429] env[70013]: DEBUG nova.compute.provider_tree [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.789742] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.816889] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52105e4e-2375-243f-fafa-ff7d996e52e1, 'name': SearchDatastore_Task, 'duration_secs': 0.014535} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.817155] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 989.817423] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] b292a985-3d2a-4cf3-a09b-8e72d8d21078/b292a985-3d2a-4cf3-a09b-8e72d8d21078.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.817986] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0981c6a9-5ba7-4074-a276-96e758afa133 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.827645] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 989.827645] env[70013]: value = "task-4231188" [ 989.827645] env[70013]: _type = "Task" [ 989.827645] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.838166] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231188, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.855908] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231186, 'name': Rename_Task, 'duration_secs': 0.202728} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.856121] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.856425] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f306abbb-acaf-43bd-b0a2-00560a85d311 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.865667] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 989.865667] env[70013]: value = "task-4231189" [ 989.865667] env[70013]: _type = "Task" [ 989.865667] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.877521] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231189, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.003567] env[70013]: DEBUG nova.network.neutron [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Updated VIF entry in instance network info cache for port a9940c14-e6da-4fa5-b093-ba598b0bdc93. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.003933] env[70013]: DEBUG nova.network.neutron [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Updating instance_info_cache with network_info: [{"id": "a9940c14-e6da-4fa5-b093-ba598b0bdc93", "address": "fa:16:3e:08:d8:08", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9940c14-e6", "ovs_interfaceid": "a9940c14-e6da-4fa5-b093-ba598b0bdc93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.076902] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231187, 'name': CreateVM_Task, 'duration_secs': 0.412407} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.077238] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.077862] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.078035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.078380] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 990.078676] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c31b1bdc-b049-4361-9293-bb65e6426e77 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.085259] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 990.085259] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52954279-bc91-26cc-315f-d61b7d0d702c" [ 990.085259] env[70013]: _type = "Task" [ 990.085259] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.097412] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52954279-bc91-26cc-315f-d61b7d0d702c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.292521] env[70013]: DEBUG nova.scheduler.client.report [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 990.296317] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "refresh_cache-69f67e8d-2da0-474c-92db-8388c98b61f4" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.296647] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 990.296993] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 990.297169] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.315971] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 990.341941] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231188, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.381392] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231189, 'name': PowerOnVM_Task} progress is 87%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.506562] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5568cf8-5cc7-4ef4-a490-1a7e55462f6b req-0d17ff9c-6123-40df-acee-57b3cca6ebf5 service nova] Releasing lock "refresh_cache-2070989d-6c90-4eb1-8508-7587ed9659d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.597745] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52954279-bc91-26cc-315f-d61b7d0d702c, 'name': SearchDatastore_Task, 'duration_secs': 0.06132} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.598066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.598354] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.598603] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.598753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.598932] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.599548] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b89aa472-52f0-4707-b0da-dae155859176 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.611526] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.611526] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.612097] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc47a037-fe7b-400c-8bb5-e5a1d9bd8de5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.617836] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 990.617836] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52be70f2-d9fb-2b26-6fe8-dc641897f002" [ 990.617836] env[70013]: _type = "Task" [ 990.617836] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.629052] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52be70f2-d9fb-2b26-6fe8-dc641897f002, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.800590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.800884] env[70013]: INFO nova.compute.manager [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Successfully reverted task state from None on failure for instance. [ 990.803556] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.227s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.803756] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.805835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.098s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.807358] env[70013]: INFO nova.compute.claims [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server [None req-1f013dab-9ff7-4ca6-8b13-08615d82106f tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server yield [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5088b04e-0b5f-4444-afcf-aa0f1b4fc257"}]} [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 990.811183] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 990.812570] env[70013]: ERROR oslo_messaging.rpc.server [ 990.821824] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.841270] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231188, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5597} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.841270] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] b292a985-3d2a-4cf3-a09b-8e72d8d21078/b292a985-3d2a-4cf3-a09b-8e72d8d21078.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 990.841270] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.841976] env[70013]: INFO nova.scheduler.client.report [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted allocations for instance 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df [ 990.842975] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18c3578b-d07f-4fef-9bf7-91191cfa24f2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.855210] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 990.855210] env[70013]: value = "task-4231190" [ 990.855210] env[70013]: _type = "Task" [ 990.855210] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.866416] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231190, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.876363] env[70013]: DEBUG oslo_vmware.api [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231189, 'name': PowerOnVM_Task, 'duration_secs': 0.893583} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.877749] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.877983] env[70013]: INFO nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Took 9.23 seconds to spawn the instance on the hypervisor. [ 990.878195] env[70013]: DEBUG nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 990.879114] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63a4f64-d0ac-4af2-a437-1326b5621023 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.130361] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52be70f2-d9fb-2b26-6fe8-dc641897f002, 'name': SearchDatastore_Task, 'duration_secs': 0.016986} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.131320] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f01fb2b-8e03-40c7-b539-58238b981a0a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.138272] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 991.138272] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e3fd2f-b932-75a2-1654-d7cde12f056d" [ 991.138272] env[70013]: _type = "Task" [ 991.138272] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.146925] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e3fd2f-b932-75a2-1654-d7cde12f056d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.324949] env[70013]: INFO nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 69f67e8d-2da0-474c-92db-8388c98b61f4] Took 1.03 seconds to deallocate network for instance. [ 991.352954] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0013a290-56f6-4381-a769-acaf8844d4d9 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.640s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 991.364337] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231190, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080296} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.364601] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.366135] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9287587-d478-435e-8f83-315566664106 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.392620] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] b292a985-3d2a-4cf3-a09b-8e72d8d21078/b292a985-3d2a-4cf3-a09b-8e72d8d21078.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.393294] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3349a742-3300-480d-b0cf-54158a33db7f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.414613] env[70013]: INFO nova.compute.manager [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Took 45.08 seconds to build instance. [ 991.418549] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 991.418549] env[70013]: value = "task-4231191" [ 991.418549] env[70013]: _type = "Task" [ 991.418549] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.427752] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.652179] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e3fd2f-b932-75a2-1654-d7cde12f056d, 'name': SearchDatastore_Task, 'duration_secs': 0.054458} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.652179] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.652179] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2070989d-6c90-4eb1-8508-7587ed9659d6/2070989d-6c90-4eb1-8508-7587ed9659d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.652179] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3206bd8-b6cf-4944-9216-452ac402b053 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.662136] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 991.662136] env[70013]: value = "task-4231192" [ 991.662136] env[70013]: _type = "Task" [ 991.662136] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.673088] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231192, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.919118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-3109eb60-e869-4959-b970-bdace95cf0ab tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.884s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 991.938621] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231191, 'name': ReconfigVM_Task, 'duration_secs': 0.316252} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.940735] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Reconfigured VM instance instance-0000004a to attach disk [datastore2] b292a985-3d2a-4cf3-a09b-8e72d8d21078/b292a985-3d2a-4cf3-a09b-8e72d8d21078.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.940735] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2155a60-70d1-4b84-8453-89eae499c136 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.947630] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 991.947630] env[70013]: value = "task-4231193" [ 991.947630] env[70013]: _type = "Task" [ 991.947630] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.981608] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231193, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.173878] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231192, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.371589] env[70013]: INFO nova.scheduler.client.report [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Deleted allocations for instance 69f67e8d-2da0-474c-92db-8388c98b61f4 [ 992.414343] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b72a3d-8602-43af-ab81-b741b6e0a33c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.426028] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e531615-f385-4d7f-b8df-ee343f29fa84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.430663] env[70013]: DEBUG nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 992.471816] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedb1f72-7755-4dd8-a102-81558a2624b3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.483746] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231193, 'name': Rename_Task, 'duration_secs': 0.245145} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.485441] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.485751] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-517dcd12-7bf8-446e-b138-bf77dc3a30dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.488320] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047afb38-79f1-43a5-bc7d-fc33aa826ba3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.504517] env[70013]: DEBUG nova.compute.provider_tree [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 992.507318] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 992.507318] env[70013]: value = "task-4231194" [ 992.507318] env[70013]: _type = "Task" [ 992.507318] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.517435] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231194, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.673093] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231192, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.711784} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.673478] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2070989d-6c90-4eb1-8508-7587ed9659d6/2070989d-6c90-4eb1-8508-7587ed9659d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.673772] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.674098] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8641c3a-84a6-4ca6-8131-e926427afc32 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.683271] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 992.683271] env[70013]: value = "task-4231195" [ 992.683271] env[70013]: _type = "Task" [ 992.683271] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.699272] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231195, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.887867] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "69f67e8d-2da0-474c-92db-8388c98b61f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.706s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.994327] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 993.027253] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231194, 'name': PowerOnVM_Task} progress is 87%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.033278] env[70013]: ERROR nova.scheduler.client.report [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [req-717161c2-65cf-4444-b84b-dcf141f6bb8d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-717161c2-65cf-4444-b84b-dcf141f6bb8d"}]} [ 993.033736] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.034463] env[70013]: ERROR nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Traceback (most recent call last): [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] yield [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] self.set_inventory_for_provider( [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-717161c2-65cf-4444-b84b-dcf141f6bb8d"}]} [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] During handling of the above exception, another exception occurred: [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Traceback (most recent call last): [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] with self.rt.instance_claim(context, instance, node, allocs, [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] return f(*args, **kwargs) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] self._update(elevated, cn) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] self._update_to_placement(context, compute_node, startup) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] return attempt.get(self._wrap_exception) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] six.reraise(self.value[0], self.value[1], self.value[2]) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] raise value [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] self.reportclient.update_from_provider_tree( [ 993.034463] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 993.035594] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] with catch_all(pd.uuid): [ 993.035594] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 993.035594] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] self.gen.throw(typ, value, traceback) [ 993.035594] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 993.035594] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] raise exception.ResourceProviderSyncFailed() [ 993.035594] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 993.035594] env[70013]: ERROR nova.compute.manager [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] [ 993.035594] env[70013]: DEBUG nova.compute.utils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 993.039595] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.170s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.039792] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.041972] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.103s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.044065] env[70013]: INFO nova.compute.claims [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.047992] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Build of instance 5a535178-36b5-4015-9cd4-1a3b9b7517c0 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 993.049233] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 993.049233] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquiring lock "refresh_cache-5a535178-36b5-4015-9cd4-1a3b9b7517c0" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.049233] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Acquired lock "refresh_cache-5a535178-36b5-4015-9cd4-1a3b9b7517c0" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.049233] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.077129] env[70013]: INFO nova.scheduler.client.report [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Deleted allocations for instance ab091ef4-9828-438f-8e95-4370bfbd34de [ 993.194362] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231195, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070424} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.195775] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.195775] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df94e26-19dc-4e0a-a9bf-aeed08fd02eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.227027] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 2070989d-6c90-4eb1-8508-7587ed9659d6/2070989d-6c90-4eb1-8508-7587ed9659d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.227027] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3c22920-acc6-495b-aca7-08e0f4f2f456 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.252919] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 993.252919] env[70013]: value = "task-4231196" [ 993.252919] env[70013]: _type = "Task" [ 993.252919] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.262513] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231196, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.526584] env[70013]: DEBUG oslo_vmware.api [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231194, 'name': PowerOnVM_Task, 'duration_secs': 0.876748} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.526881] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.528812] env[70013]: INFO nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Took 8.76 seconds to spawn the instance on the hypervisor. [ 993.528812] env[70013]: DEBUG nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 993.528812] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2f22d9-23b8-4991-9b35-5c6e2e92bd39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.590217] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e22046a-0628-4a09-9720-edfadf6a3427 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "ab091ef4-9828-438f-8e95-4370bfbd34de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.736s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.661898] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.748100] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.766518] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231196, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.059670] env[70013]: INFO nova.compute.manager [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Took 45.27 seconds to build instance. [ 994.089041] env[70013]: DEBUG nova.scheduler.client.report [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 994.108226] env[70013]: DEBUG nova.scheduler.client.report [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 994.108226] env[70013]: DEBUG nova.compute.provider_tree [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 994.128762] env[70013]: DEBUG nova.scheduler.client.report [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 994.152595] env[70013]: DEBUG nova.scheduler.client.report [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 994.251554] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Releasing lock "refresh_cache-5a535178-36b5-4015-9cd4-1a3b9b7517c0" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.251554] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 994.251926] env[70013]: DEBUG nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 994.251926] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 994.265455] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231196, 'name': ReconfigVM_Task, 'duration_secs': 0.850353} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.268908] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 2070989d-6c90-4eb1-8508-7587ed9659d6/2070989d-6c90-4eb1-8508-7587ed9659d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.269780] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd9e58a8-3ae1-44cf-a5f0-52a8393e9098 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.286791] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 994.286791] env[70013]: value = "task-4231200" [ 994.286791] env[70013]: _type = "Task" [ 994.286791] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.300312] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231200, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.309541] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 994.565777] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c036ac33-a915-48a8-9adc-00c6fbd12462 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.281s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.657718] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.658045] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 994.658323] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.658547] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 994.658766] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.663913] env[70013]: INFO nova.compute.manager [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Terminating instance [ 994.687786] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c783669-5135-43e1-a995-c545544440fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.698887] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0b9ee6-3033-487f-ab8a-6a1d25e935d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.734286] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e78dd2-301f-45ac-bc7c-c3190ea60f82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.743522] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962a9464-519d-4b09-8ba3-10836b5338f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.762079] env[70013]: DEBUG nova.compute.provider_tree [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 994.792222] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231200, 'name': Rename_Task, 'duration_secs': 0.240523} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.792519] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.792875] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9126abd-9d69-4a55-a8a2-9bc284bfae99 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.799699] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 994.799699] env[70013]: value = "task-4231201" [ 994.799699] env[70013]: _type = "Task" [ 994.799699] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.808043] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.813011] env[70013]: DEBUG nova.network.neutron [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.904036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "54d39dfd-6429-4009-8d26-22c1fa46243e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.904036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 994.904036] env[70013]: DEBUG nova.compute.manager [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 994.904036] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12636604-79f5-457c-ba7a-d57bbb4ff75e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.911865] env[70013]: DEBUG nova.compute.manager [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 994.912563] env[70013]: DEBUG nova.objects.instance [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lazy-loading 'flavor' on Instance uuid 54d39dfd-6429-4009-8d26-22c1fa46243e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.167732] env[70013]: DEBUG nova.compute.manager [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 995.167979] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 995.169042] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3889d8e1-78e2-42da-a18d-f5d5fbdf39b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.178285] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 995.178554] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c6cff3e-3bb8-4c99-b6c4-82f2c66e8faa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.185670] env[70013]: DEBUG oslo_vmware.api [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 995.185670] env[70013]: value = "task-4231202" [ 995.185670] env[70013]: _type = "Task" [ 995.185670] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.194431] env[70013]: DEBUG oslo_vmware.api [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231202, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.289844] env[70013]: ERROR nova.scheduler.client.report [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [req-2b930171-19c4-468d-aa48-d046b52dcc90] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2b930171-19c4-468d-aa48-d046b52dcc90"}]} [ 995.290738] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.291981] env[70013]: ERROR nova.compute.manager [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Traceback (most recent call last): [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] yield [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] self.set_inventory_for_provider( [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2b930171-19c4-468d-aa48-d046b52dcc90"}]} [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] During handling of the above exception, another exception occurred: [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Traceback (most recent call last): [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] with self.rt.instance_claim(context, instance, node, allocs, [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] return f(*args, **kwargs) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] self._update(elevated, cn) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] self._update_to_placement(context, compute_node, startup) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] return attempt.get(self._wrap_exception) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] six.reraise(self.value[0], self.value[1], self.value[2]) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] raise value [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] self.reportclient.update_from_provider_tree( [ 995.291981] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 995.292871] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] with catch_all(pd.uuid): [ 995.292871] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 995.292871] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] self.gen.throw(typ, value, traceback) [ 995.292871] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 995.292871] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] raise exception.ResourceProviderSyncFailed() [ 995.292871] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 995.292871] env[70013]: ERROR nova.compute.manager [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] [ 995.293537] env[70013]: DEBUG nova.compute.utils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 995.295810] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.762s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.296279] env[70013]: DEBUG nova.objects.instance [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'resources' on Instance uuid 30fcb8c7-0947-426e-9aa5-081446e31c28 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.298762] env[70013]: DEBUG nova.compute.manager [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Build of instance 81e32169-6f75-4a62-9605-9c84dae7d417 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 995.299531] env[70013]: DEBUG nova.compute.manager [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 995.299995] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "refresh_cache-81e32169-6f75-4a62-9605-9c84dae7d417" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.300326] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "refresh_cache-81e32169-6f75-4a62-9605-9c84dae7d417" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.300663] env[70013]: DEBUG nova.network.neutron [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.315815] env[70013]: INFO nova.compute.manager [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] [instance: 5a535178-36b5-4015-9cd4-1a3b9b7517c0] Took 1.06 seconds to deallocate network for instance. [ 995.318879] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231201, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.697062] env[70013]: DEBUG oslo_vmware.api [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231202, 'name': PowerOffVM_Task, 'duration_secs': 0.19579} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.697062] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.697062] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 995.697321] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6357458-2ed0-4010-bb24-bcfdc8f51ada {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.767031] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.767384] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.767633] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Deleting the datastore file [datastore2] b292a985-3d2a-4cf3-a09b-8e72d8d21078 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.767925] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1773cd4-0fc0-4158-898c-a96e01712752 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.780019] env[70013]: DEBUG oslo_vmware.api [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for the task: (returnval){ [ 995.780019] env[70013]: value = "task-4231204" [ 995.780019] env[70013]: _type = "Task" [ 995.780019] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.789114] env[70013]: DEBUG oslo_vmware.api [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231204, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.800339] env[70013]: DEBUG nova.objects.instance [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'numa_topology' on Instance uuid 30fcb8c7-0947-426e-9aa5-081446e31c28 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.814276] env[70013]: DEBUG oslo_vmware.api [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231201, 'name': PowerOnVM_Task, 'duration_secs': 0.781727} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.815546] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.815900] env[70013]: INFO nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Took 8.89 seconds to spawn the instance on the hypervisor. [ 995.818254] env[70013]: DEBUG nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 995.818254] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48d4ae7-80a3-42e3-8811-2733ea331f68 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.834335] env[70013]: DEBUG nova.network.neutron [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 995.922718] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 995.924034] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b4405bd-2417-4507-a48a-53dfde2d8b0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.932813] env[70013]: DEBUG nova.network.neutron [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.938314] env[70013]: DEBUG oslo_vmware.api [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 995.938314] env[70013]: value = "task-4231205" [ 995.938314] env[70013]: _type = "Task" [ 995.938314] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.949748] env[70013]: DEBUG oslo_vmware.api [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231205, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.291146] env[70013]: DEBUG oslo_vmware.api [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Task: {'id': task-4231204, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169688} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.291508] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.291594] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.291785] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.292078] env[70013]: INFO nova.compute.manager [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Took 1.12 seconds to destroy the instance on the hypervisor. [ 996.292416] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 996.292627] env[70013]: DEBUG nova.compute.manager [-] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 996.292725] env[70013]: DEBUG nova.network.neutron [-] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.302552] env[70013]: DEBUG nova.objects.base [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Object Instance<30fcb8c7-0947-426e-9aa5-081446e31c28> lazy-loaded attributes: resources,numa_topology {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 996.319794] env[70013]: DEBUG nova.scheduler.client.report [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 996.335197] env[70013]: DEBUG nova.scheduler.client.report [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 996.335438] env[70013]: DEBUG nova.compute.provider_tree [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.348216] env[70013]: INFO nova.compute.manager [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Took 45.30 seconds to build instance. [ 996.351454] env[70013]: INFO nova.scheduler.client.report [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Deleted allocations for instance 5a535178-36b5-4015-9cd4-1a3b9b7517c0 [ 996.358873] env[70013]: DEBUG nova.scheduler.client.report [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 996.382762] env[70013]: DEBUG nova.scheduler.client.report [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 996.439614] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "refresh_cache-81e32169-6f75-4a62-9605-9c84dae7d417" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.439875] env[70013]: DEBUG nova.compute.manager [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 996.440062] env[70013]: DEBUG nova.compute.manager [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 996.440241] env[70013]: DEBUG nova.network.neutron [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.456239] env[70013]: DEBUG oslo_vmware.api [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231205, 'name': PowerOffVM_Task, 'duration_secs': 0.23468} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.459723] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 996.459723] env[70013]: DEBUG nova.compute.manager [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 996.459723] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8186be-0490-401d-805d-1c943bbb55cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.478515] env[70013]: DEBUG nova.network.neutron [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 996.496340] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "181238eb-cb0c-4740-9896-ca745b53ebe8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.496340] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.767865] env[70013]: DEBUG nova.compute.manager [req-c847ed3b-29df-4721-b3ba-6162e95f83a3 req-0c53e316-c59e-406c-88d0-db6048c9104c service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Received event network-vif-deleted-ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 996.767865] env[70013]: INFO nova.compute.manager [req-c847ed3b-29df-4721-b3ba-6162e95f83a3 req-0c53e316-c59e-406c-88d0-db6048c9104c service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Neutron deleted interface ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7; detaching it from the instance and deleting it from the info cache [ 996.767865] env[70013]: DEBUG nova.network.neutron [req-c847ed3b-29df-4721-b3ba-6162e95f83a3 req-0c53e316-c59e-406c-88d0-db6048c9104c service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.851098] env[70013]: DEBUG oslo_concurrency.lockutils [None req-952887b7-5920-4d76-ac0a-1b55ec6c6663 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.422s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.866349] env[70013]: DEBUG oslo_concurrency.lockutils [None req-70197690-36d0-4950-ae9c-984fffc7b6a9 tempest-MultipleCreateTestJSON-373751725 tempest-MultipleCreateTestJSON-373751725-project-member] Lock "5a535178-36b5-4015-9cd4-1a3b9b7517c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.647s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.873098] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea2ae96-3868-497f-b211-16a85da02b65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.881022] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704426dd-257f-46d6-973c-0f002755ba03 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.915628] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff55768-c9c7-4e9f-b955-cb7fbc309efc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.923777] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccaa5a90-05ee-44d6-b835-b20afb47dc56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.941262] env[70013]: DEBUG nova.compute.provider_tree [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.983485] env[70013]: DEBUG nova.network.neutron [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.984529] env[70013]: DEBUG oslo_concurrency.lockutils [None req-31089511-7584-4fbf-8346-a0fd39ab9356 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.082s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.003304] env[70013]: DEBUG nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 997.182875] env[70013]: DEBUG nova.network.neutron [-] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.183893] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.184016] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.268943] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-205dd37d-1d3e-4a92-9869-be4e2444ba57 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.281066] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bab5aa6-a3ab-4e73-b9d8-9fe2150c4b82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.317804] env[70013]: DEBUG nova.compute.manager [req-c847ed3b-29df-4721-b3ba-6162e95f83a3 req-0c53e316-c59e-406c-88d0-db6048c9104c service nova] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Detach interface failed, port_id=ccf0fb0e-eb83-4463-acc1-fc5bab8a5fb7, reason: Instance b292a985-3d2a-4cf3-a09b-8e72d8d21078 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 997.462472] env[70013]: ERROR nova.scheduler.client.report [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [req-71e7993e-732d-49b9-9e07-f5733cc11272] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-71e7993e-732d-49b9-9e07-f5733cc11272"}]} [ 997.462889] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.167s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.463087] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 59.179s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.465041] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.863s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.465041] env[70013]: DEBUG nova.objects.instance [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lazy-loading 'resources' on Instance uuid ef64a05a-b514-4c35-81d3-664ae1ad3ff1 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.466105] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 36.597s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.466330] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.466597] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.466697] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.469124] env[70013]: INFO nova.compute.manager [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Terminating instance [ 997.487189] env[70013]: INFO nova.compute.manager [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 81e32169-6f75-4a62-9605-9c84dae7d417] Took 1.05 seconds to deallocate network for instance. [ 997.537248] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.647635] env[70013]: DEBUG nova.objects.instance [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lazy-loading 'flavor' on Instance uuid 54d39dfd-6429-4009-8d26-22c1fa46243e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.687064] env[70013]: INFO nova.compute.manager [-] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Took 1.39 seconds to deallocate network for instance. [ 997.687627] env[70013]: DEBUG nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 997.973235] env[70013]: DEBUG nova.compute.manager [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 997.973538] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.980157] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f6c82811-1ff2-4710-883c-a24920a340ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.993577] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a6a8f2-c730-4b68-85bc-ca57d41ad71f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.008167] env[70013]: DEBUG nova.scheduler.client.report [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 998.036085] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 30fcb8c7-0947-426e-9aa5-081446e31c28 could not be found. [ 998.036365] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.036498] env[70013]: INFO nova.compute.manager [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Took 0.06 seconds to destroy the instance on the hypervisor. [ 998.037396] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 998.037835] env[70013]: DEBUG nova.scheduler.client.report [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 998.038033] env[70013]: DEBUG nova.compute.provider_tree [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 998.040226] env[70013]: DEBUG nova.compute.manager [-] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 998.041859] env[70013]: DEBUG nova.network.neutron [-] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 998.053836] env[70013]: DEBUG nova.scheduler.client.report [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 998.077439] env[70013]: DEBUG nova.scheduler.client.report [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 998.153573] env[70013]: DEBUG oslo_concurrency.lockutils [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.153741] env[70013]: DEBUG oslo_concurrency.lockutils [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquired lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 998.154419] env[70013]: DEBUG nova.network.neutron [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 998.154575] env[70013]: DEBUG nova.objects.instance [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lazy-loading 'info_cache' on Instance uuid 54d39dfd-6429-4009-8d26-22c1fa46243e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.202169] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.219470] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.546127] env[70013]: INFO nova.scheduler.client.report [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted allocations for instance 81e32169-6f75-4a62-9605-9c84dae7d417 [ 998.571476] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b20abb3-fad8-47e8-9823-4caff726c052 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.582134] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f7c48b-04d8-4b34-b0ce-df019a455023 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.616454] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a89803-4e3c-4275-ae2f-36267035d913 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.625570] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67efd77-4e1d-4a4b-bc5a-d62784ba8090 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.642464] env[70013]: DEBUG nova.compute.provider_tree [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 998.661399] env[70013]: DEBUG nova.objects.base [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Object Instance<54d39dfd-6429-4009-8d26-22c1fa46243e> lazy-loaded attributes: flavor,info_cache {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 998.797181] env[70013]: DEBUG nova.compute.manager [req-e8f97977-c2ea-42f3-bc1e-362e23a1ac1c req-96b040c8-a1a1-4b5c-8689-efc45be48ace service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Received event network-vif-deleted-72c3d938-1cba-4e01-8176-aa8e8a245c4c {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 998.797181] env[70013]: INFO nova.compute.manager [req-e8f97977-c2ea-42f3-bc1e-362e23a1ac1c req-96b040c8-a1a1-4b5c-8689-efc45be48ace service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Neutron deleted interface 72c3d938-1cba-4e01-8176-aa8e8a245c4c; detaching it from the instance and deleting it from the info cache [ 998.797181] env[70013]: DEBUG nova.network.neutron [req-e8f97977-c2ea-42f3-bc1e-362e23a1ac1c req-96b040c8-a1a1-4b5c-8689-efc45be48ace service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.824415] env[70013]: DEBUG nova.network.neutron [-] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.994682] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.058545] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7973b4c6-f5ff-429d-945c-758282ebb735 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "81e32169-6f75-4a62-9605-9c84dae7d417" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.681s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.169971] env[70013]: ERROR nova.scheduler.client.report [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [req-ca0d279d-62f0-455b-8f2c-a6644c055a3c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ca0d279d-62f0-455b-8f2c-a6644c055a3c"}]} [ 999.170476] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.171209] env[70013]: ERROR nova.compute.manager [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Traceback (most recent call last): [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] yield [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self.set_inventory_for_provider( [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ca0d279d-62f0-455b-8f2c-a6644c055a3c"}]} [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] During handling of the above exception, another exception occurred: [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Traceback (most recent call last): [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self._delete_instance(context, instance, bdms) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self._complete_deletion(context, instance) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self._update_resource_tracker(context, instance) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self.rt.update_usage(context, instance, instance.node) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] return f(*args, **kwargs) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self._update(context.elevated(), self.compute_nodes[nodename]) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self._update_to_placement(context, compute_node, startup) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] return attempt.get(self._wrap_exception) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] six.reraise(self.value[0], self.value[1], self.value[2]) [ 999.171209] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] raise value [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self.reportclient.update_from_provider_tree( [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] with catch_all(pd.uuid): [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] self.gen.throw(typ, value, traceback) [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] raise exception.ResourceProviderSyncFailed() [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 999.172904] env[70013]: ERROR nova.compute.manager [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] [ 999.177667] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.965s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.177919] env[70013]: DEBUG nova.objects.instance [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lazy-loading 'resources' on Instance uuid 39557c50-866a-4510-b840-b1a6a3e3890e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.300378] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ee56af3-394a-4119-a00f-884ab9ddb2a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.312995] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ccbf03-39cf-4b31-8a29-a732809c5f3c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.328301] env[70013]: INFO nova.compute.manager [-] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Took 1.29 seconds to deallocate network for instance. [ 999.358625] env[70013]: DEBUG nova.compute.manager [req-e8f97977-c2ea-42f3-bc1e-362e23a1ac1c req-96b040c8-a1a1-4b5c-8689-efc45be48ace service nova] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Detach interface failed, port_id=72c3d938-1cba-4e01-8176-aa8e8a245c4c, reason: Instance 30fcb8c7-0947-426e-9aa5-081446e31c28 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 999.494348] env[70013]: DEBUG nova.network.neutron [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Updating instance_info_cache with network_info: [{"id": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "address": "fa:16:3e:1b:aa:bf", "network": {"id": "66a20513-27d1-46c4-8749-18416964efd1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1620950280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c45f05e8334dcca93a5c6d4ef4c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20f2f376-c1", "ovs_interfaceid": "20f2f376-c1c7-4171-bc55-1ecdb0cc5d14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.684317] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.508s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.706805] env[70013]: DEBUG nova.scheduler.client.report [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 999.724031] env[70013]: DEBUG nova.scheduler.client.report [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 999.724031] env[70013]: DEBUG nova.compute.provider_tree [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.741020] env[70013]: DEBUG nova.scheduler.client.report [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 999.762061] env[70013]: DEBUG nova.scheduler.client.report [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 999.839801] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.997379] env[70013]: DEBUG oslo_concurrency.lockutils [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Releasing lock "refresh_cache-54d39dfd-6429-4009-8d26-22c1fa46243e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1000.173144] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d44a43-9b09-4c44-9f93-9895c8b63039 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.181750] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93238768-169d-40da-85a7-d729eb55f99f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.223083] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d36b499-f4e7-4c8b-a1f4-08defa3312e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.230017] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826f9ae3-1dd2-40aa-a5c7-52c8ec896335 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.247533] env[70013]: DEBUG nova.compute.provider_tree [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.610272] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "1ff7f11d-360f-44dd-9ca7-f76c1db905fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.610488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "1ff7f11d-360f-44dd-9ca7-f76c1db905fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.725076] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.769494] env[70013]: ERROR nova.scheduler.client.report [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [req-9855243a-70dd-434c-846b-338e0d72c52a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9855243a-70dd-434c-846b-338e0d72c52a"}]} [ 1000.769788] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.592s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.770436] env[70013]: ERROR nova.compute.manager [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Traceback (most recent call last): [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] yield [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self.set_inventory_for_provider( [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9855243a-70dd-434c-846b-338e0d72c52a"}]} [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] During handling of the above exception, another exception occurred: [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Traceback (most recent call last): [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self._delete_instance(context, instance, bdms) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self._complete_deletion(context, instance) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self._update_resource_tracker(context, instance) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self.rt.update_usage(context, instance, instance.node) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] return f(*args, **kwargs) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self._update_to_placement(context, compute_node, startup) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] return attempt.get(self._wrap_exception) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1000.770436] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] raise value [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self.reportclient.update_from_provider_tree( [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] with catch_all(pd.uuid): [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] self.gen.throw(typ, value, traceback) [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] raise exception.ResourceProviderSyncFailed() [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1000.771620] env[70013]: ERROR nova.compute.manager [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] [ 1000.772959] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.368s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.774394] env[70013]: INFO nova.compute.claims [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1000.865068] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "789fffd9-6725-4bf6-9144-dd603b0a521f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.865068] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.865427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "789fffd9-6725-4bf6-9144-dd603b0a521f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.865427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.865568] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.868146] env[70013]: INFO nova.compute.manager [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Terminating instance [ 1001.003244] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.004491] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-273d4152-c285-439e-8f4d-9fc549d10fcc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.011897] env[70013]: DEBUG oslo_vmware.api [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 1001.011897] env[70013]: value = "task-4231209" [ 1001.011897] env[70013]: _type = "Task" [ 1001.011897] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.026835] env[70013]: DEBUG oslo_vmware.api [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231209, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.116592] env[70013]: DEBUG nova.compute.manager [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1001.278682] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "39557c50-866a-4510-b840-b1a6a3e3890e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.111s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.373309] env[70013]: DEBUG nova.compute.manager [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1001.373560] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.374469] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3465ff-ff19-4816-83a8-1b0d5fc37986 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.383258] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.383484] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1695e2f-9460-4527-a68f-ee7d1adece5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.390974] env[70013]: DEBUG oslo_vmware.api [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 1001.390974] env[70013]: value = "task-4231210" [ 1001.390974] env[70013]: _type = "Task" [ 1001.390974] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.399933] env[70013]: DEBUG oslo_vmware.api [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.524635] env[70013]: DEBUG oslo_vmware.api [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231209, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.638609] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1001.807111] env[70013]: DEBUG nova.scheduler.client.report [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1001.821777] env[70013]: DEBUG nova.scheduler.client.report [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1001.822048] env[70013]: DEBUG nova.compute.provider_tree [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1001.835497] env[70013]: DEBUG nova.scheduler.client.report [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1001.856834] env[70013]: DEBUG nova.scheduler.client.report [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1001.901867] env[70013]: DEBUG oslo_vmware.api [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231210, 'name': PowerOffVM_Task, 'duration_secs': 0.229328} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.904942] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.905147] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.905854] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6514ea60-9a75-4ae3-88b3-e24300107b46 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.975090] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.975090] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.975090] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Deleting the datastore file [datastore1] 789fffd9-6725-4bf6-9144-dd603b0a521f {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.975090] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c6014a6-2102-4699-9e7a-e7ea557fb3e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.982852] env[70013]: DEBUG oslo_vmware.api [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for the task: (returnval){ [ 1001.982852] env[70013]: value = "task-4231212" [ 1001.982852] env[70013]: _type = "Task" [ 1001.982852] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.992924] env[70013]: DEBUG oslo_vmware.api [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231212, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.025686] env[70013]: DEBUG oslo_vmware.api [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231209, 'name': PowerOnVM_Task, 'duration_secs': 0.585766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.029766] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.030229] env[70013]: DEBUG nova.compute.manager [None req-760bc055-c778-4b88-895d-286eb09f1c09 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1002.032423] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7616f5-0786-4fe8-99d6-2ffd93f5e0fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.348680] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0839813f-869a-4b6e-aa7c-060e93279a45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.356674] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18205ed-a401-4b0c-a23f-0c7bff045c60 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.390761] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd427c0-8ff5-44b9-b6f4-c47ebbd00dfd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.398981] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353c67b9-36ab-4424-a8a6-8bc59fc1b6e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.414360] env[70013]: DEBUG nova.compute.provider_tree [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1002.494715] env[70013]: DEBUG oslo_vmware.api [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Task: {'id': task-4231212, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234953} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.494996] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.495199] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.495376] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.495549] env[70013]: INFO nova.compute.manager [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1002.495794] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1002.496055] env[70013]: DEBUG nova.compute.manager [-] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1002.496161] env[70013]: DEBUG nova.network.neutron [-] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.795746] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.877098] env[70013]: DEBUG nova.compute.manager [req-57b2892d-cede-4e2f-839d-f4614da53313 req-8fd4329f-990b-49ec-8a9c-88c0154cac47 service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Received event network-vif-deleted-4a9242b0-4afa-443e-9ccd-dc35313fc59a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1002.877377] env[70013]: INFO nova.compute.manager [req-57b2892d-cede-4e2f-839d-f4614da53313 req-8fd4329f-990b-49ec-8a9c-88c0154cac47 service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Neutron deleted interface 4a9242b0-4afa-443e-9ccd-dc35313fc59a; detaching it from the instance and deleting it from the info cache [ 1002.877436] env[70013]: DEBUG nova.network.neutron [req-57b2892d-cede-4e2f-839d-f4614da53313 req-8fd4329f-990b-49ec-8a9c-88c0154cac47 service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.947220] env[70013]: ERROR nova.scheduler.client.report [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [req-b4898201-310c-4c72-bf68-ed36e8654ba9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b4898201-310c-4c72-bf68-ed36e8654ba9"}]} [ 1002.947610] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.175s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.948236] env[70013]: ERROR nova.compute.manager [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Traceback (most recent call last): [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] yield [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] self.set_inventory_for_provider( [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b4898201-310c-4c72-bf68-ed36e8654ba9"}]} [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] During handling of the above exception, another exception occurred: [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Traceback (most recent call last): [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] with self.rt.instance_claim(context, instance, node, allocs, [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] return f(*args, **kwargs) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] self._update(elevated, cn) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] self._update_to_placement(context, compute_node, startup) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] return attempt.get(self._wrap_exception) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] raise value [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] self.reportclient.update_from_provider_tree( [ 1002.948236] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1002.949627] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] with catch_all(pd.uuid): [ 1002.949627] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1002.949627] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] self.gen.throw(typ, value, traceback) [ 1002.949627] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1002.949627] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] raise exception.ResourceProviderSyncFailed() [ 1002.949627] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1002.949627] env[70013]: ERROR nova.compute.manager [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] [ 1002.949627] env[70013]: DEBUG nova.compute.utils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1002.950303] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.756s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.950529] env[70013]: DEBUG nova.objects.instance [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lazy-loading 'resources' on Instance uuid 8dbef727-13e9-4231-95a8-65d015ee13be {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.952039] env[70013]: DEBUG nova.compute.manager [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Build of instance a8bb5ed8-00cc-493d-9ab4-264d878d4135 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1002.952478] env[70013]: DEBUG nova.compute.manager [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1002.952716] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Acquiring lock "refresh_cache-a8bb5ed8-00cc-493d-9ab4-264d878d4135" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.952858] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Acquired lock "refresh_cache-a8bb5ed8-00cc-493d-9ab4-264d878d4135" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1002.953355] env[70013]: DEBUG nova.network.neutron [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1003.286023] env[70013]: DEBUG nova.network.neutron [-] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.381129] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-843fa6ac-e583-4bc5-b876-17e1bda996b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.393911] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe096e6e-e207-436f-b9af-0eab48beb1e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.434403] env[70013]: DEBUG nova.compute.manager [req-57b2892d-cede-4e2f-839d-f4614da53313 req-8fd4329f-990b-49ec-8a9c-88c0154cac47 service nova] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Detach interface failed, port_id=4a9242b0-4afa-443e-9ccd-dc35313fc59a, reason: Instance 789fffd9-6725-4bf6-9144-dd603b0a521f could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1003.477945] env[70013]: DEBUG nova.scheduler.client.report [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1003.481490] env[70013]: DEBUG nova.network.neutron [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1003.510622] env[70013]: DEBUG nova.scheduler.client.report [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1003.510622] env[70013]: DEBUG nova.compute.provider_tree [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1003.527280] env[70013]: DEBUG nova.scheduler.client.report [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1003.556449] env[70013]: DEBUG nova.scheduler.client.report [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1003.572170] env[70013]: DEBUG nova.network.neutron [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.787225] env[70013]: INFO nova.compute.manager [-] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Took 1.29 seconds to deallocate network for instance. [ 1004.017632] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254348f8-f498-4b33-bf1d-cc8c1ec7cd1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.027279] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f69e86-77e0-4bc4-bd80-03010181c287 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.068622] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178d69a5-ecbc-490c-af8b-005e97b159cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.076689] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Releasing lock "refresh_cache-a8bb5ed8-00cc-493d-9ab4-264d878d4135" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1004.076935] env[70013]: DEBUG nova.compute.manager [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1004.077137] env[70013]: DEBUG nova.compute.manager [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1004.077456] env[70013]: DEBUG nova.network.neutron [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.080855] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb59289e-4316-4f01-b574-ca62cc40469c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.098139] env[70013]: DEBUG nova.compute.provider_tree [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.100965] env[70013]: DEBUG nova.network.neutron [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1004.297755] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.604021] env[70013]: DEBUG nova.network.neutron [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.625342] env[70013]: ERROR nova.scheduler.client.report [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [req-e35264e5-8c42-4c01-a1ec-ad8e2017dcb2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e35264e5-8c42-4c01-a1ec-ad8e2017dcb2"}]} [ 1004.625682] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.626311] env[70013]: ERROR nova.compute.manager [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Traceback (most recent call last): [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] yield [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self.set_inventory_for_provider( [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e35264e5-8c42-4c01-a1ec-ad8e2017dcb2"}]} [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] During handling of the above exception, another exception occurred: [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Traceback (most recent call last): [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self._delete_instance(context, instance, bdms) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self._complete_deletion(context, instance) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self._update_resource_tracker(context, instance) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self.rt.update_usage(context, instance, instance.node) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] return f(*args, **kwargs) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self._update_to_placement(context, compute_node, startup) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] return attempt.get(self._wrap_exception) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1004.626311] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] raise value [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self.reportclient.update_from_provider_tree( [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] with catch_all(pd.uuid): [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] self.gen.throw(typ, value, traceback) [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] raise exception.ResourceProviderSyncFailed() [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1004.627455] env[70013]: ERROR nova.compute.manager [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] [ 1004.628809] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.807s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.630379] env[70013]: INFO nova.compute.claims [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.109997] env[70013]: INFO nova.compute.manager [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] [instance: a8bb5ed8-00cc-493d-9ab4-264d878d4135] Took 1.03 seconds to deallocate network for instance. [ 1005.139801] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.453s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.664776] env[70013]: DEBUG nova.scheduler.client.report [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1005.680320] env[70013]: DEBUG nova.scheduler.client.report [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1005.680559] env[70013]: DEBUG nova.compute.provider_tree [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.692387] env[70013]: DEBUG nova.scheduler.client.report [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1005.708039] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.708313] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.708520] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "9a61dedd-3764-4bd9-a300-480cc7d14a21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.708704] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.708870] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.710649] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "eab0c393-4d3d-4659-8225-c82ddadb330b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.710878] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.711089] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.711277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.711452] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.713521] env[70013]: DEBUG nova.scheduler.client.report [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1005.715760] env[70013]: INFO nova.compute.manager [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Terminating instance [ 1005.717353] env[70013]: INFO nova.compute.manager [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Terminating instance [ 1006.079543] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9a3ad0-d6be-4324-9867-77203032d864 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.088039] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c02b64-8299-4966-ba5d-5b6efeffa5eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.899027] env[70013]: DEBUG nova.compute.manager [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1006.899027] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.899027] env[70013]: DEBUG nova.compute.manager [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1006.899027] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.899632] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110773fe-cda6-4384-b156-097cc1e76822 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.902803] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7be429-15f7-4554-b629-fa867db29b81 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.909213] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4e943a-5b38-4b0c-972b-3d00b085a34f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.911432] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "74f39e88-b034-4e45-ae1c-453d644c093b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.911903] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "74f39e88-b034-4e45-ae1c-453d644c093b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.917877] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.922615] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f91f3d1d-eefe-4182-b601-e0c45e6e739e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.924368] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.925440] env[70013]: INFO nova.scheduler.client.report [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Deleted allocations for instance a8bb5ed8-00cc-493d-9ab4-264d878d4135 [ 1006.930618] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c40a2c1-3bea-4d98-9e3c-908d6402ab7d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.937024] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af14d427-48f3-4520-99c4-0dd8e627cc9c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.945637] env[70013]: DEBUG oslo_vmware.api [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 1006.945637] env[70013]: value = "task-4231216" [ 1006.945637] env[70013]: _type = "Task" [ 1006.945637] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.954797] env[70013]: DEBUG nova.compute.provider_tree [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.961042] env[70013]: DEBUG oslo_vmware.api [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 1006.961042] env[70013]: value = "task-4231215" [ 1006.961042] env[70013]: _type = "Task" [ 1006.961042] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.968387] env[70013]: DEBUG oslo_vmware.api [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231216, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.973632] env[70013]: DEBUG oslo_vmware.api [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231215, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.401031] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.415974] env[70013]: DEBUG nova.compute.manager [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1007.442219] env[70013]: DEBUG oslo_concurrency.lockutils [None req-73c2081c-944e-4d27-a50a-130129aae140 tempest-FloatingIPsAssociationTestJSON-1659716564 tempest-FloatingIPsAssociationTestJSON-1659716564-project-member] Lock "a8bb5ed8-00cc-493d-9ab4-264d878d4135" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.663s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.474670] env[70013]: DEBUG oslo_vmware.api [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231216, 'name': PowerOffVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.483928] env[70013]: DEBUG oslo_vmware.api [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231215, 'name': PowerOffVM_Task, 'duration_secs': 0.529922} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.484041] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.484365] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.485270] env[70013]: ERROR nova.scheduler.client.report [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [req-9ad29e61-630e-4e2d-977d-3283f0b3c3cb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9ad29e61-630e-4e2d-977d-3283f0b3c3cb"}]} [ 1007.485621] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.857s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.486305] env[70013]: ERROR nova.compute.manager [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] Traceback (most recent call last): [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] yield [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] self.set_inventory_for_provider( [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9ad29e61-630e-4e2d-977d-3283f0b3c3cb"}]} [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] During handling of the above exception, another exception occurred: [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] Traceback (most recent call last): [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] with self.rt.instance_claim(context, instance, node, allocs, [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] return f(*args, **kwargs) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] self._update(elevated, cn) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] self._update_to_placement(context, compute_node, startup) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] return attempt.get(self._wrap_exception) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] raise value [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] self.reportclient.update_from_provider_tree( [ 1007.486305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1007.487305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] with catch_all(pd.uuid): [ 1007.487305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1007.487305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] self.gen.throw(typ, value, traceback) [ 1007.487305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1007.487305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] raise exception.ResourceProviderSyncFailed() [ 1007.487305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1007.487305] env[70013]: ERROR nova.compute.manager [instance: 5c855194-feef-403f-9279-a32d5492c353] [ 1007.487305] env[70013]: DEBUG nova.compute.utils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1007.488111] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d564523-d4df-4aa5-8bad-37c3a783076a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.489920] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.470s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.491498] env[70013]: INFO nova.compute.claims [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.495476] env[70013]: DEBUG nova.compute.manager [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Build of instance 5c855194-feef-403f-9279-a32d5492c353 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1007.495825] env[70013]: DEBUG nova.compute.manager [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1007.496066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "refresh_cache-5c855194-feef-403f-9279-a32d5492c353" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.496213] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired lock "refresh_cache-5c855194-feef-403f-9279-a32d5492c353" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.496370] env[70013]: DEBUG nova.network.neutron [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.615594] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.615594] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.615594] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Deleting the datastore file [datastore2] 9a61dedd-3764-4bd9-a300-480cc7d14a21 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.615594] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ddf3afb-1ed2-4346-b474-820e2d140ed5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.623260] env[70013]: DEBUG oslo_vmware.api [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 1007.623260] env[70013]: value = "task-4231219" [ 1007.623260] env[70013]: _type = "Task" [ 1007.623260] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.632063] env[70013]: DEBUG oslo_vmware.api [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.731784] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.731784] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.732175] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.732175] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.732353] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.735175] env[70013]: INFO nova.compute.manager [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Terminating instance [ 1007.942338] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.963711] env[70013]: DEBUG oslo_vmware.api [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231216, 'name': PowerOffVM_Task, 'duration_secs': 0.529428} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.964020] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.964195] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.964624] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4293d79a-24b1-47af-a09c-d49dd5aa1315 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.024091] env[70013]: DEBUG nova.network.neutron [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.030278] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.030510] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.030742] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleting the datastore file [datastore1] eab0c393-4d3d-4659-8225-c82ddadb330b {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.031907] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c56f674-2c8c-445e-9ba7-a5f60db2c7de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.038525] env[70013]: DEBUG oslo_vmware.api [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for the task: (returnval){ [ 1008.038525] env[70013]: value = "task-4231221" [ 1008.038525] env[70013]: _type = "Task" [ 1008.038525] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.048283] env[70013]: DEBUG oslo_vmware.api [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231221, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.102456] env[70013]: DEBUG nova.network.neutron [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.134323] env[70013]: DEBUG oslo_vmware.api [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231219, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153032} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.134579] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.138018] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.138018] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.138018] env[70013]: INFO nova.compute.manager [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1008.138018] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1008.138018] env[70013]: DEBUG nova.compute.manager [-] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1008.138018] env[70013]: DEBUG nova.network.neutron [-] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.241224] env[70013]: DEBUG nova.compute.manager [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1008.241475] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.242460] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a571034-a913-4f5a-a1eb-ebcf3b465802 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.252338] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.252746] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c781c929-a348-4ec9-bb2a-fa984df529d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.260089] env[70013]: DEBUG oslo_vmware.api [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 1008.260089] env[70013]: value = "task-4231222" [ 1008.260089] env[70013]: _type = "Task" [ 1008.260089] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.270570] env[70013]: DEBUG oslo_vmware.api [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4231222, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.428419] env[70013]: DEBUG nova.compute.manager [req-da15dcd8-0339-4702-a7ea-9520a113bacf req-a7623649-fe2e-4b5a-bc60-19fae396654f service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Received event network-vif-deleted-e7899a16-e277-4a87-8992-3900bdc7c327 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1008.428629] env[70013]: INFO nova.compute.manager [req-da15dcd8-0339-4702-a7ea-9520a113bacf req-a7623649-fe2e-4b5a-bc60-19fae396654f service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Neutron deleted interface e7899a16-e277-4a87-8992-3900bdc7c327; detaching it from the instance and deleting it from the info cache [ 1008.428835] env[70013]: DEBUG nova.network.neutron [req-da15dcd8-0339-4702-a7ea-9520a113bacf req-a7623649-fe2e-4b5a-bc60-19fae396654f service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.522841] env[70013]: DEBUG nova.scheduler.client.report [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1008.540120] env[70013]: DEBUG nova.scheduler.client.report [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1008.540706] env[70013]: DEBUG nova.compute.provider_tree [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.554824] env[70013]: DEBUG oslo_vmware.api [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Task: {'id': task-4231221, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.355214} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.556032] env[70013]: DEBUG nova.scheduler.client.report [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1008.558308] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.558621] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.558788] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.559065] env[70013]: INFO nova.compute.manager [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1008.559309] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1008.559805] env[70013]: DEBUG nova.compute.manager [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1008.559905] env[70013]: DEBUG nova.network.neutron [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.584645] env[70013]: DEBUG nova.scheduler.client.report [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1008.606243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Releasing lock "refresh_cache-5c855194-feef-403f-9279-a32d5492c353" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.606243] env[70013]: DEBUG nova.compute.manager [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1008.606592] env[70013]: DEBUG nova.compute.manager [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1008.606812] env[70013]: DEBUG nova.network.neutron [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.643698] env[70013]: DEBUG nova.network.neutron [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.771152] env[70013]: DEBUG oslo_vmware.api [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4231222, 'name': PowerOffVM_Task, 'duration_secs': 0.348857} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.781726] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.781726] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.781726] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5abfcf1-ecb0-44a5-b4d4-dfeb9cc1edd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.854252] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.856039] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.856039] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Deleting the datastore file [datastore1] 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.856039] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2cd8ce59-14cd-4180-9a53-87c7c37ee4e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.870472] env[70013]: DEBUG oslo_vmware.api [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for the task: (returnval){ [ 1008.870472] env[70013]: value = "task-4231224" [ 1008.870472] env[70013]: _type = "Task" [ 1008.870472] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.879188] env[70013]: DEBUG nova.network.neutron [-] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.889762] env[70013]: DEBUG oslo_vmware.api [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4231224, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.933912] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd4ac0ad-168d-4923-b6cd-caf97bd3e109 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.942251] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52400994-70a7-454f-8cdd-7c0f03434832 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.975809] env[70013]: DEBUG nova.compute.manager [req-da15dcd8-0339-4702-a7ea-9520a113bacf req-a7623649-fe2e-4b5a-bc60-19fae396654f service nova] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Detach interface failed, port_id=e7899a16-e277-4a87-8992-3900bdc7c327, reason: Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1009.101093] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3c5371-eccc-45c0-aab2-78e6b4f74bb9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.107661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7437e8-72b3-4521-92c8-144eb23f4641 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.143754] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e461588-7399-454d-ab54-30a721652f09 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.146547] env[70013]: DEBUG nova.network.neutron [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.153370] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14770ea1-e740-41d6-a000-8d3b27ad87ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.168958] env[70013]: DEBUG nova.compute.provider_tree [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.384337] env[70013]: INFO nova.compute.manager [-] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Took 1.25 seconds to deallocate network for instance. [ 1009.384868] env[70013]: DEBUG oslo_vmware.api [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Task: {'id': task-4231224, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149929} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.386393] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.386580] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.386753] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.386923] env[70013]: INFO nova.compute.manager [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1009.387172] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1009.389987] env[70013]: DEBUG nova.compute.manager [-] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1009.390153] env[70013]: DEBUG nova.network.neutron [-] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.393894] env[70013]: DEBUG nova.network.neutron [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.649817] env[70013]: INFO nova.compute.manager [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5c855194-feef-403f-9279-a32d5492c353] Took 1.04 seconds to deallocate network for instance. [ 1009.672592] env[70013]: DEBUG nova.scheduler.client.report [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1009.893487] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.905682] env[70013]: INFO nova.compute.manager [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Took 1.35 seconds to deallocate network for instance. [ 1010.183183] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.693s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.183806] env[70013]: DEBUG nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1010.186543] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.811s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.186825] env[70013]: DEBUG nova.objects.instance [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lazy-loading 'resources' on Instance uuid 85488f5e-bd8a-4bcf-8420-744c54176c2e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.413107] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.461591] env[70013]: DEBUG nova.compute.manager [req-6d8fca26-2f32-4db6-bee9-58c68072f5f8 req-25013e9b-7f7c-41c3-a0f8-bf5a3013ac56 service nova] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Received event network-vif-deleted-f777b44c-5cf6-4ccb-86d9-15dad465d158 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1010.461887] env[70013]: DEBUG nova.compute.manager [req-6d8fca26-2f32-4db6-bee9-58c68072f5f8 req-25013e9b-7f7c-41c3-a0f8-bf5a3013ac56 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Received event network-vif-deleted-04205eac-90e1-47c0-82c4-703b9258de68 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1010.462051] env[70013]: INFO nova.compute.manager [req-6d8fca26-2f32-4db6-bee9-58c68072f5f8 req-25013e9b-7f7c-41c3-a0f8-bf5a3013ac56 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Neutron deleted interface 04205eac-90e1-47c0-82c4-703b9258de68; detaching it from the instance and deleting it from the info cache [ 1010.462227] env[70013]: DEBUG nova.network.neutron [req-6d8fca26-2f32-4db6-bee9-58c68072f5f8 req-25013e9b-7f7c-41c3-a0f8-bf5a3013ac56 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.491984] env[70013]: DEBUG nova.network.neutron [-] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.690575] env[70013]: DEBUG nova.compute.utils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1010.697147] env[70013]: INFO nova.scheduler.client.report [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Deleted allocations for instance 5c855194-feef-403f-9279-a32d5492c353 [ 1010.702731] env[70013]: DEBUG nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1010.967412] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e548af7d-89ee-4be4-bb66-20c8b3dbcf4f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.980857] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d246da11-26fd-4e95-9c9f-5a6a8beeb9e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.996706] env[70013]: INFO nova.compute.manager [-] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Took 1.61 seconds to deallocate network for instance. [ 1011.017249] env[70013]: DEBUG nova.compute.manager [req-6d8fca26-2f32-4db6-bee9-58c68072f5f8 req-25013e9b-7f7c-41c3-a0f8-bf5a3013ac56 service nova] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Detach interface failed, port_id=04205eac-90e1-47c0-82c4-703b9258de68, reason: Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1011.154891] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc3a439-114d-4e3f-b725-f94e6c684c38 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.163941] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226bca45-ca9e-4dd7-84f4-ec2481e2e067 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.197534] env[70013]: DEBUG nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1011.201555] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830799ba-71ac-49cc-8d62-148de96360ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.204813] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ecbf1b82-181e-46bb-8288-bd45b2c9e2d9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "5c855194-feef-403f-9279-a32d5492c353" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.858s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.210485] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cae029-1029-47a5-9a0a-c89646844be7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.226781] env[70013]: DEBUG nova.compute.provider_tree [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.504350] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.731378] env[70013]: DEBUG nova.scheduler.client.report [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.216450] env[70013]: DEBUG nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1012.237279] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.050s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1012.239277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.603s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.241925] env[70013]: INFO nova.compute.claims [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.248164] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1012.248164] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.248164] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1012.248164] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.248620] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1012.248836] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1012.249072] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1012.249242] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1012.249410] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1012.249569] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1012.249988] env[70013]: DEBUG nova.virt.hardware [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1012.253042] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5560c4f-dbff-49f1-adb5-d4e3d1059560 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.260079] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4b033d-86d4-4fa3-8b1b-057745ae13b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.267715] env[70013]: INFO nova.scheduler.client.report [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Deleted allocations for instance 85488f5e-bd8a-4bcf-8420-744c54176c2e [ 1012.285025] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.291348] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Creating folder: Project (ceb8cfcae49f4d4287a3f4c0e2c88962). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.294538] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e2588657-17e2-4857-9f6b-44a5cf7600a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.306113] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Created folder: Project (ceb8cfcae49f4d4287a3f4c0e2c88962) in parent group-v836999. [ 1012.306334] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Creating folder: Instances. Parent ref: group-v837179. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.306584] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a6c6f0b-4533-46a2-9b34-bcac5f6607f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.318257] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Created folder: Instances in parent group-v837179. [ 1012.321109] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1012.321109] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1012.321109] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3978052e-0847-48fb-98e5-c3b1a5d6cdec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.340451] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.340451] env[70013]: value = "task-4231228" [ 1012.340451] env[70013]: _type = "Task" [ 1012.340451] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.349470] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231228, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.795351] env[70013]: DEBUG oslo_concurrency.lockutils [None req-19c38719-a6c5-48f3-b8bc-74b2b2850aff tempest-DeleteServersAdminTestJSON-1843386794 tempest-DeleteServersAdminTestJSON-1843386794-project-member] Lock "85488f5e-bd8a-4bcf-8420-744c54176c2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.625s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1012.854070] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231228, 'name': CreateVM_Task, 'duration_secs': 0.345481} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.854798] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1012.854798] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.854967] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1012.855334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1012.855636] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa4967b4-7a86-4931-a06a-9ed410e18418 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.861160] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1012.861160] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c64886-7f05-b1ea-26a6-5a2d32ef19d8" [ 1012.861160] env[70013]: _type = "Task" [ 1012.861160] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.872070] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c64886-7f05-b1ea-26a6-5a2d32ef19d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.372107] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c64886-7f05-b1ea-26a6-5a2d32ef19d8, 'name': SearchDatastore_Task, 'duration_secs': 0.016202} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.374736] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1013.374982] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.375261] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.375408] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.375582] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.376934] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b74f2cd-b0a5-4697-94e9-b927fb45a714 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.387387] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.387643] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.388489] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89325e5a-f484-4ff3-a0d1-b15dacb1cce8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.397453] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1013.397453] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52882cc2-cc9f-9244-cdc3-64500d45db63" [ 1013.397453] env[70013]: _type = "Task" [ 1013.397453] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.407266] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52882cc2-cc9f-9244-cdc3-64500d45db63, 'name': SearchDatastore_Task, 'duration_secs': 0.010289} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.410648] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c73266fb-2dad-4570-b472-ce09bcf2561b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.416304] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1013.416304] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dce291-6cca-4d07-5571-de09316e83cb" [ 1013.416304] env[70013]: _type = "Task" [ 1013.416304] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.426584] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dce291-6cca-4d07-5571-de09316e83cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.683299] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32457e3b-0fe8-4aa4-9eb3-2bbfb97bf6ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.692464] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12044940-dec3-4775-b176-0db93f3a6f4f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.727611] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635690dd-7ed7-4b66-8fe7-444ee311712b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.734070] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "5661914f-ccd6-48da-bccc-a6b9065338c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.734070] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "5661914f-ccd6-48da-bccc-a6b9065338c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.741738] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14478e06-85fc-4cae-a633-9d23b5f66a37 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.758840] env[70013]: DEBUG nova.compute.provider_tree [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.764271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Acquiring lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.764667] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.931109] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dce291-6cca-4d07-5571-de09316e83cb, 'name': SearchDatastore_Task, 'duration_secs': 0.009852} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.931109] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1013.931109] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1013.931109] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79da30a5-1484-46e9-b175-6d0a183fd9bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.938021] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1013.938021] env[70013]: value = "task-4231229" [ 1013.938021] env[70013]: _type = "Task" [ 1013.938021] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.944744] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231229, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.236421] env[70013]: DEBUG nova.compute.manager [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1014.267229] env[70013]: DEBUG nova.scheduler.client.report [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1014.270301] env[70013]: DEBUG nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1014.447777] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231229, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492939} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.449344] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1014.449344] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1014.449344] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b96db479-46f8-49bc-be8a-f8b97f048ef7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.458031] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1014.458031] env[70013]: value = "task-4231230" [ 1014.458031] env[70013]: _type = "Task" [ 1014.458031] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.466502] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231230, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.762348] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.776639] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.777195] env[70013]: DEBUG nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1014.782192] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.788s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.783788] env[70013]: INFO nova.compute.claims [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.800590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.967153] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231230, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071344} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.967390] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.968194] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ca5d37-f82a-4f2f-93e7-0e4ecf650463 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.990397] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.990397] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-609a6ea7-dbf0-4947-a71d-844d5ee46468 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.011194] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1015.011194] env[70013]: value = "task-4231231" [ 1015.011194] env[70013]: _type = "Task" [ 1015.011194] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.019605] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231231, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.284742] env[70013]: DEBUG nova.compute.utils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1015.286229] env[70013]: DEBUG nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1015.286438] env[70013]: DEBUG nova.network.neutron [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1015.328602] env[70013]: DEBUG nova.policy [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88d1be071a1c47d8b7664380ae025351', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68ad346a8e3a465da70d3a7de825ac6d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1015.526358] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231231, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.615560] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.615832] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.617633] env[70013]: DEBUG nova.network.neutron [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Successfully created port: 4bbb7861-dce4-4ba4-9a50-78e66aa137ef {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.791382] env[70013]: DEBUG nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1016.024799] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231231, 'name': ReconfigVM_Task, 'duration_secs': 0.583824} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.025111] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.025732] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-41082a1b-6c18-4e04-bde9-b9063b8c5bcc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.034716] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1016.034716] env[70013]: value = "task-4231232" [ 1016.034716] env[70013]: _type = "Task" [ 1016.034716] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.043895] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231232, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.122094] env[70013]: DEBUG nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1016.224527] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6071d5-f686-4ec2-90e1-c7de1a8c2308 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.233076] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d68558-998e-4bfa-9404-e797f552d3c8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.266688] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179ac76a-8f72-460f-84b8-d8685ea8871c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.275834] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5b2759-0418-4742-ab3d-258baf24f71e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.291267] env[70013]: DEBUG nova.compute.provider_tree [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.545830] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231232, 'name': Rename_Task, 'duration_secs': 0.141014} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.546180] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1016.546273] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c584c1a1-b006-4a78-b200-e673a1d3baa3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.554923] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1016.554923] env[70013]: value = "task-4231233" [ 1016.554923] env[70013]: _type = "Task" [ 1016.554923] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.566774] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.643805] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.795825] env[70013]: DEBUG nova.scheduler.client.report [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1016.801940] env[70013]: DEBUG nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1016.829865] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1016.830401] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.830683] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1016.831026] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.831232] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1016.831426] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1016.831649] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1016.831859] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1016.832111] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1016.833062] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1016.833062] env[70013]: DEBUG nova.virt.hardware [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1016.833889] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4bf206e-87fc-4d41-a304-e8e0ef7686af {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.843669] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa69ed0-73ae-4559-b8dd-1fc4d4e9475f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.066279] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231233, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.068545] env[70013]: DEBUG nova.compute.manager [req-41c1a7e0-b96e-4341-bbc6-70ba4241dcb5 req-f3eef29a-feca-4f49-af44-1a2b97b9ff65 service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Received event network-vif-plugged-4bbb7861-dce4-4ba4-9a50-78e66aa137ef {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1017.068767] env[70013]: DEBUG oslo_concurrency.lockutils [req-41c1a7e0-b96e-4341-bbc6-70ba4241dcb5 req-f3eef29a-feca-4f49-af44-1a2b97b9ff65 service nova] Acquiring lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1017.069055] env[70013]: DEBUG oslo_concurrency.lockutils [req-41c1a7e0-b96e-4341-bbc6-70ba4241dcb5 req-f3eef29a-feca-4f49-af44-1a2b97b9ff65 service nova] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1017.069251] env[70013]: DEBUG oslo_concurrency.lockutils [req-41c1a7e0-b96e-4341-bbc6-70ba4241dcb5 req-f3eef29a-feca-4f49-af44-1a2b97b9ff65 service nova] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.069429] env[70013]: DEBUG nova.compute.manager [req-41c1a7e0-b96e-4341-bbc6-70ba4241dcb5 req-f3eef29a-feca-4f49-af44-1a2b97b9ff65 service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] No waiting events found dispatching network-vif-plugged-4bbb7861-dce4-4ba4-9a50-78e66aa137ef {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1017.069594] env[70013]: WARNING nova.compute.manager [req-41c1a7e0-b96e-4341-bbc6-70ba4241dcb5 req-f3eef29a-feca-4f49-af44-1a2b97b9ff65 service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Received unexpected event network-vif-plugged-4bbb7861-dce4-4ba4-9a50-78e66aa137ef for instance with vm_state building and task_state spawning. [ 1017.177697] env[70013]: DEBUG nova.network.neutron [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Successfully updated port: 4bbb7861-dce4-4ba4-9a50-78e66aa137ef {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1017.300808] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.301476] env[70013]: DEBUG nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1017.304710] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.768s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1017.306105] env[70013]: INFO nova.compute.claims [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.567223] env[70013]: DEBUG oslo_vmware.api [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231233, 'name': PowerOnVM_Task, 'duration_secs': 0.79385} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.567545] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.567719] env[70013]: INFO nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Took 5.35 seconds to spawn the instance on the hypervisor. [ 1017.567950] env[70013]: DEBUG nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1017.568706] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0d4a62-7a8b-42f0-b2fa-6ee5e6d4f5fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.819235] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "refresh_cache-073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.819235] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "refresh_cache-073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.819235] env[70013]: DEBUG nova.network.neutron [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.819235] env[70013]: DEBUG nova.compute.utils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1017.819235] env[70013]: DEBUG nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1017.819235] env[70013]: DEBUG nova.network.neutron [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1017.883230] env[70013]: DEBUG nova.policy [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d6c00fc04b8435388ae5e2652f8e0eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51b2e9e21dcc447d86aa3d2f5f307d4f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1018.091514] env[70013]: INFO nova.compute.manager [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Took 32.09 seconds to build instance. [ 1018.188479] env[70013]: DEBUG nova.network.neutron [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Successfully created port: 3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.239639] env[70013]: DEBUG nova.network.neutron [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1018.319031] env[70013]: DEBUG nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1018.471307] env[70013]: DEBUG nova.network.neutron [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Updating instance_info_cache with network_info: [{"id": "4bbb7861-dce4-4ba4-9a50-78e66aa137ef", "address": "fa:16:3e:27:48:eb", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bbb7861-dc", "ovs_interfaceid": "4bbb7861-dce4-4ba4-9a50-78e66aa137ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.594905] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a66ffcb9-cf11-4289-935e-1b80ebfb69a0 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.314s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.769028] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35af9e47-2806-4957-953e-cc845f7bc9aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.777955] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078d8a1b-9a6e-4396-9467-c2327dcf2eca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.809732] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c84f863-5607-414d-a6ec-73c1d82080d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.818475] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0be3832-5a5e-41a8-b34e-0ed0b6e1299e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.835412] env[70013]: DEBUG nova.compute.provider_tree [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.975707] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "refresh_cache-073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.976123] env[70013]: DEBUG nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Instance network_info: |[{"id": "4bbb7861-dce4-4ba4-9a50-78e66aa137ef", "address": "fa:16:3e:27:48:eb", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bbb7861-dc", "ovs_interfaceid": "4bbb7861-dce4-4ba4-9a50-78e66aa137ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1018.976631] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:48:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4bbb7861-dce4-4ba4-9a50-78e66aa137ef', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.986492] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1018.986797] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1018.987110] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd9b678f-bb6d-4d1a-bf8f-5bb8ba4d1244 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.004675] env[70013]: INFO nova.compute.manager [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Rebuilding instance [ 1019.013833] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.013833] env[70013]: value = "task-4231234" [ 1019.013833] env[70013]: _type = "Task" [ 1019.013833] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.024025] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231234, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.051973] env[70013]: DEBUG nova.compute.manager [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1019.052990] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27b69e3-e2b5-4ce6-8934-eea001b992e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.150281] env[70013]: DEBUG nova.compute.manager [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Received event network-changed-4bbb7861-dce4-4ba4-9a50-78e66aa137ef {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1019.151130] env[70013]: DEBUG nova.compute.manager [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Refreshing instance network info cache due to event network-changed-4bbb7861-dce4-4ba4-9a50-78e66aa137ef. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1019.151762] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] Acquiring lock "refresh_cache-073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.151762] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] Acquired lock "refresh_cache-073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1019.151762] env[70013]: DEBUG nova.network.neutron [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Refreshing network info cache for port 4bbb7861-dce4-4ba4-9a50-78e66aa137ef {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.338957] env[70013]: DEBUG nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1019.341906] env[70013]: DEBUG nova.scheduler.client.report [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1019.367167] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1019.367479] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.367659] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1019.367927] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.368138] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1019.368344] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1019.368606] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1019.368788] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1019.369189] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1019.369404] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1019.369591] env[70013]: DEBUG nova.virt.hardware [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1019.370694] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c749a08f-abf4-4ea9-a3bd-4a64f959e3b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.379727] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69b4e1a-1964-4537-a825-7b104178a7c7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.524329] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231234, 'name': CreateVM_Task, 'duration_secs': 0.341469} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.524503] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1019.525167] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.525344] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1019.525697] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1019.525961] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f8df791-fe8e-4048-a5ec-0a31f473b42e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.530498] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1019.530498] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526b20bc-6e6c-e732-62fd-afc6b69eaafc" [ 1019.530498] env[70013]: _type = "Task" [ 1019.530498] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.538395] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526b20bc-6e6c-e732-62fd-afc6b69eaafc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.851907] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.853339] env[70013]: DEBUG nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1019.858246] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.656s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.858769] env[70013]: DEBUG nova.objects.instance [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lazy-loading 'resources' on Instance uuid b292a985-3d2a-4cf3-a09b-8e72d8d21078 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.970871] env[70013]: DEBUG nova.network.neutron [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Successfully updated port: 3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1019.973918] env[70013]: DEBUG nova.network.neutron [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Updated VIF entry in instance network info cache for port 4bbb7861-dce4-4ba4-9a50-78e66aa137ef. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.974296] env[70013]: DEBUG nova.network.neutron [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Updating instance_info_cache with network_info: [{"id": "4bbb7861-dce4-4ba4-9a50-78e66aa137ef", "address": "fa:16:3e:27:48:eb", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bbb7861-dc", "ovs_interfaceid": "4bbb7861-dce4-4ba4-9a50-78e66aa137ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.042138] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526b20bc-6e6c-e732-62fd-afc6b69eaafc, 'name': SearchDatastore_Task, 'duration_secs': 0.01122} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.042477] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.042719] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.042957] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.043122] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.043413] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.043951] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53f73d91-12ae-4e1a-809d-fcc536a1d141 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.054976] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.055182] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1020.055957] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffb4fd10-cbc8-480f-abd6-a222efc7d6e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.061724] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1020.061724] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c467b3-8fbf-a9f9-f410-8b2d3b848698" [ 1020.061724] env[70013]: _type = "Task" [ 1020.061724] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.075192] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.075192] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c467b3-8fbf-a9f9-f410-8b2d3b848698, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.075306] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-177c0992-90a2-44dd-8683-73ba2bf7c4c6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.083807] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1020.083807] env[70013]: value = "task-4231235" [ 1020.083807] env[70013]: _type = "Task" [ 1020.083807] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.093417] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231235, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.365930] env[70013]: DEBUG nova.compute.utils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1020.367575] env[70013]: DEBUG nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1020.367772] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1020.409105] env[70013]: DEBUG nova.policy [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '804d90ab34784afe864a6b76379619bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4eef90bcd1794b788b1340d14c6df552', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1020.477943] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d4953c5-9ae7-4208-8c91-7b386abb553f req-659a7bc9-0dfb-45fd-bbd4-ad2024e9e65a service nova] Releasing lock "refresh_cache-073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.478628] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.478782] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.478985] env[70013]: DEBUG nova.network.neutron [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.575593] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c467b3-8fbf-a9f9-f410-8b2d3b848698, 'name': SearchDatastore_Task, 'duration_secs': 0.023016} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.579853] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44ca9d69-5214-452a-8c8a-8b537d14ec53 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.589653] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1020.589653] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c65ee5-00ea-0114-e562-2c3b3fa25244" [ 1020.589653] env[70013]: _type = "Task" [ 1020.589653] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.600021] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231235, 'name': PowerOffVM_Task, 'duration_secs': 0.129213} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.601428] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.601600] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.602369] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddfe89d-b2f1-48c2-9115-dd917f944bd6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.608966] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c65ee5-00ea-0114-e562-2c3b3fa25244, 'name': SearchDatastore_Task, 'duration_secs': 0.010346} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.609745] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.609944] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7/073f3ebf-1672-47fc-b4ed-cdb3651a8ee7.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1020.610222] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b393a42-0457-4da4-9a0d-c9cee67675e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.616167] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1020.620116] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24dd2310-cf6e-40eb-b379-1bfa47ebf6c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.625113] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1020.625113] env[70013]: value = "task-4231236" [ 1020.625113] env[70013]: _type = "Task" [ 1020.625113] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.634368] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.658393] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1020.658784] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1020.659096] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Deleting the datastore file [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.659463] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ed3d975-42aa-407e-b903-f16c87ecfdc6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.668822] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1020.668822] env[70013]: value = "task-4231238" [ 1020.668822] env[70013]: _type = "Task" [ 1020.668822] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.683039] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.722714] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Successfully created port: ebcb7b3a-3c04-471a-b585-af59dfa574f3 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.851316] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e28b51-8bec-4066-b9d3-feee4f7d60c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.858884] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69b344c-e848-4926-b8db-c0e5dee82af8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.892254] env[70013]: DEBUG nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1020.895494] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d9a81e-a1d2-46af-9734-c5906a634cb0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.907225] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4142a9-8964-4a23-8b6f-3c105925bfd1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.930850] env[70013]: DEBUG nova.compute.provider_tree [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.035976] env[70013]: DEBUG nova.network.neutron [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1021.133587] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Successfully created port: 96be8274-aa70-44e5-8a0c-84ad7d0f9737 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1021.140232] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231236, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.184909] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108625} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.185075] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.185288] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.185461] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.250632] env[70013]: DEBUG nova.compute.manager [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Received event network-vif-plugged-3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1021.250846] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] Acquiring lock "1230ca35-1512-4464-85fc-d3b4ab05eac1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1021.251197] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.251361] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.251532] env[70013]: DEBUG nova.compute.manager [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] No waiting events found dispatching network-vif-plugged-3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1021.251724] env[70013]: WARNING nova.compute.manager [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Received unexpected event network-vif-plugged-3d1c82dc-c975-4156-8d68-368f68ec3f87 for instance with vm_state building and task_state spawning. [ 1021.251894] env[70013]: DEBUG nova.compute.manager [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Received event network-changed-3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1021.253951] env[70013]: DEBUG nova.compute.manager [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Refreshing instance network info cache due to event network-changed-3d1c82dc-c975-4156-8d68-368f68ec3f87. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1021.254214] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] Acquiring lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.292876] env[70013]: DEBUG nova.network.neutron [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Updating instance_info_cache with network_info: [{"id": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "address": "fa:16:3e:d3:e4:4b", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d1c82dc-c9", "ovs_interfaceid": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.436662] env[70013]: DEBUG nova.scheduler.client.report [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1021.638087] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231236, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533207} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.638353] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7/073f3ebf-1672-47fc-b4ed-cdb3651a8ee7.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1021.638573] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.638838] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6421161-3984-4fcf-b74b-8c34a85ca2e9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.645960] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1021.645960] env[70013]: value = "task-4231239" [ 1021.645960] env[70013]: _type = "Task" [ 1021.645960] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.657171] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231239, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.796323] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.796590] env[70013]: DEBUG nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Instance network_info: |[{"id": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "address": "fa:16:3e:d3:e4:4b", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d1c82dc-c9", "ovs_interfaceid": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1021.797310] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] Acquired lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.797496] env[70013]: DEBUG nova.network.neutron [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Refreshing network info cache for port 3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.799060] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:e4:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'abcf0d10-3f3f-45dc-923e-1c78766e2dad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d1c82dc-c975-4156-8d68-368f68ec3f87', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1021.807267] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1021.810609] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1021.811147] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e16f659-9d0f-478f-b28d-d495520bec8b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.833610] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1021.833610] env[70013]: value = "task-4231240" [ 1021.833610] env[70013]: _type = "Task" [ 1021.833610] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.844347] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231240, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.909914] env[70013]: DEBUG nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1021.941106] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1021.941365] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.941525] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1021.941707] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.941861] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1021.942019] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1021.942392] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1021.942392] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1021.942536] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1021.943198] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1021.943198] env[70013]: DEBUG nova.virt.hardware [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1021.943629] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.949139] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b93edf4-cf14-4417-95f9-10afc0444a49 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.949721] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.730s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.951284] env[70013]: INFO nova.compute.claims [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.963967] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92efa6f-fcbf-4462-8bfb-886dcdb67163 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.983319] env[70013]: INFO nova.scheduler.client.report [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Deleted allocations for instance b292a985-3d2a-4cf3-a09b-8e72d8d21078 [ 1022.076537] env[70013]: DEBUG nova.network.neutron [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Updated VIF entry in instance network info cache for port 3d1c82dc-c975-4156-8d68-368f68ec3f87. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1022.077152] env[70013]: DEBUG nova.network.neutron [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Updating instance_info_cache with network_info: [{"id": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "address": "fa:16:3e:d3:e4:4b", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d1c82dc-c9", "ovs_interfaceid": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.157358] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231239, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067833} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.157713] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.159066] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632d017a-fbf1-4ffe-ae05-13fc47333c10 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.186147] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7/073f3ebf-1672-47fc-b4ed-cdb3651a8ee7.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.186508] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b2cc2bd-784c-45ef-a0d7-a12487691882 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.210774] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1022.210774] env[70013]: value = "task-4231241" [ 1022.210774] env[70013]: _type = "Task" [ 1022.210774] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.222275] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231241, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.231245] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1022.231638] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.231882] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1022.232157] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.232384] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1022.232602] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1022.232883] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1022.233187] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1022.233418] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1022.233650] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1022.233843] env[70013]: DEBUG nova.virt.hardware [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1022.234871] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ece483f-559c-4f84-8710-ec10d07894db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.246516] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edc5cd8-b1bd-43b1-9f52-5fc3042b265d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.261803] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.267504] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1022.267795] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.268080] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73e3773d-ff08-4692-8491-2a106da51513 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.286935] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.286935] env[70013]: value = "task-4231242" [ 1022.286935] env[70013]: _type = "Task" [ 1022.286935] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.295658] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231242, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.347275] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231240, 'name': CreateVM_Task, 'duration_secs': 0.387396} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.347511] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1022.348555] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.348724] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.349052] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1022.349369] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a189a64-a097-4e54-98ab-a93241d2baed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.356074] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1022.356074] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e88ef5-1026-1226-f767-8bc545fbcdce" [ 1022.356074] env[70013]: _type = "Task" [ 1022.356074] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.365388] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e88ef5-1026-1226-f767-8bc545fbcdce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.492161] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a47fcc18-d775-4861-8b0c-86bcf001d2f6 tempest-ServerGroupTestJSON-1018463365 tempest-ServerGroupTestJSON-1018463365-project-member] Lock "b292a985-3d2a-4cf3-a09b-8e72d8d21078" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.834s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.579453] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9d3e3db-dd1e-413a-94de-f09855fccd15 req-d7cb4b73-4de1-4fec-81ce-3d6e4a88bc92 service nova] Releasing lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.720972] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231241, 'name': ReconfigVM_Task, 'duration_secs': 0.291894} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.721405] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7/073f3ebf-1672-47fc-b4ed-cdb3651a8ee7.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.722103] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dee3de32-81bf-4170-b674-2cf788fb37ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.729386] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1022.729386] env[70013]: value = "task-4231243" [ 1022.729386] env[70013]: _type = "Task" [ 1022.729386] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.733566] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Successfully updated port: ebcb7b3a-3c04-471a-b585-af59dfa574f3 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.740297] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231243, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.798022] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231242, 'name': CreateVM_Task, 'duration_secs': 0.295142} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.798294] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1022.798751] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.867636] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e88ef5-1026-1226-f767-8bc545fbcdce, 'name': SearchDatastore_Task, 'duration_secs': 0.016092} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.868049] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.868361] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.868671] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.868889] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.869154] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.869507] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.869879] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1022.870178] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ade17d1a-eab6-4933-8619-5339dfa1fb92 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.872592] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ed33b95-65a0-4a7b-b355-d50fd61ca031 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.878185] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1022.878185] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c10a9-e6b5-5fc1-8e12-5f57466b4ce9" [ 1022.878185] env[70013]: _type = "Task" [ 1022.878185] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.882730] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.882958] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1022.884085] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48896a58-bde4-476d-8bba-7dd62c444e6a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.889394] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c10a9-e6b5-5fc1-8e12-5f57466b4ce9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.893262] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1022.893262] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527971bd-c114-b2bb-f747-8112ee5884ce" [ 1022.893262] env[70013]: _type = "Task" [ 1022.893262] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.901457] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527971bd-c114-b2bb-f747-8112ee5884ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.245597] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231243, 'name': Rename_Task, 'duration_secs': 0.16554} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.247958] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.249031] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a90a335-e763-45a9-b224-a2e41e1c6ba1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.259868] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1023.259868] env[70013]: value = "task-4231244" [ 1023.259868] env[70013]: _type = "Task" [ 1023.259868] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.272906] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231244, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.285727] env[70013]: DEBUG nova.compute.manager [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received event network-vif-plugged-ebcb7b3a-3c04-471a-b585-af59dfa574f3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1023.286139] env[70013]: DEBUG oslo_concurrency.lockutils [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] Acquiring lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.286462] env[70013]: DEBUG oslo_concurrency.lockutils [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.286743] env[70013]: DEBUG oslo_concurrency.lockutils [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.287408] env[70013]: DEBUG nova.compute.manager [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] No waiting events found dispatching network-vif-plugged-ebcb7b3a-3c04-471a-b585-af59dfa574f3 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1023.291244] env[70013]: WARNING nova.compute.manager [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received unexpected event network-vif-plugged-ebcb7b3a-3c04-471a-b585-af59dfa574f3 for instance with vm_state building and task_state spawning. [ 1023.291470] env[70013]: DEBUG nova.compute.manager [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received event network-changed-ebcb7b3a-3c04-471a-b585-af59dfa574f3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1023.291671] env[70013]: DEBUG nova.compute.manager [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Refreshing instance network info cache due to event network-changed-ebcb7b3a-3c04-471a-b585-af59dfa574f3. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1023.291908] env[70013]: DEBUG oslo_concurrency.lockutils [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] Acquiring lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.292064] env[70013]: DEBUG oslo_concurrency.lockutils [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] Acquired lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.292230] env[70013]: DEBUG nova.network.neutron [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Refreshing network info cache for port ebcb7b3a-3c04-471a-b585-af59dfa574f3 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1023.387584] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060577ed-b3e9-4f4f-9833-d7ab85952af0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.395890] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c10a9-e6b5-5fc1-8e12-5f57466b4ce9, 'name': SearchDatastore_Task, 'duration_secs': 0.252265} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.400865] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.401201] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.401425] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.404789] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e25bfad-e539-4cd3-984e-96ee4bcaeac7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.413057] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527971bd-c114-b2bb-f747-8112ee5884ce, 'name': SearchDatastore_Task, 'duration_secs': 0.238247} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.414493] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61d40cd8-1eca-4aca-94f7-3ff056c30447 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.453163] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af8fc8e-95a5-4168-a479-490eb4fc88d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.457784] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1023.457784] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5223cb40-ff82-9761-d637-3e8a5d6146dd" [ 1023.457784] env[70013]: _type = "Task" [ 1023.457784] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.465267] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cabf066-f584-4e99-9169-00de8d675ac7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.472777] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5223cb40-ff82-9761-d637-3e8a5d6146dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010697} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.473473] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.473783] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1023.474076] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.474283] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.474512] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68f6ebfe-b294-45af-980d-38e8ac2aa2f4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.484436] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1be5f0c9-ec1a-4b35-a71d-c97329207293 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.486513] env[70013]: DEBUG nova.compute.provider_tree [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.494372] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1023.494372] env[70013]: value = "task-4231245" [ 1023.494372] env[70013]: _type = "Task" [ 1023.494372] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.495638] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.495814] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.500142] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40247cf5-5c7f-4a4d-80ca-64f55841b8d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.509978] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231245, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.513874] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1023.513874] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523d34d8-b2ef-ed69-7a16-3db9397d7d80" [ 1023.513874] env[70013]: _type = "Task" [ 1023.513874] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.524775] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523d34d8-b2ef-ed69-7a16-3db9397d7d80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.777608] env[70013]: DEBUG oslo_vmware.api [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231244, 'name': PowerOnVM_Task, 'duration_secs': 0.51286} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.778067] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.778320] env[70013]: INFO nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Took 6.98 seconds to spawn the instance on the hypervisor. [ 1023.778583] env[70013]: DEBUG nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1023.779727] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89e1115-408b-40a1-84c8-862314596842 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.856676] env[70013]: DEBUG nova.network.neutron [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1023.989858] env[70013]: DEBUG nova.scheduler.client.report [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1024.002801] env[70013]: DEBUG nova.network.neutron [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.010424] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231245, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.026107] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523d34d8-b2ef-ed69-7a16-3db9397d7d80, 'name': SearchDatastore_Task, 'duration_secs': 0.010654} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.027038] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bc496ca-971b-407f-a8a1-d12a68e71bb2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.034665] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1024.034665] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52aa885d-76f2-fc89-dcd5-9b8120ed43e0" [ 1024.034665] env[70013]: _type = "Task" [ 1024.034665] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.044707] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52aa885d-76f2-fc89-dcd5-9b8120ed43e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.301024] env[70013]: INFO nova.compute.manager [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Took 34.69 seconds to build instance. [ 1024.497970] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.498567] env[70013]: DEBUG nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1024.501713] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.507s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.501713] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.501713] env[70013]: INFO nova.compute.manager [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Successfully reverted task state from None on failure for instance. [ 1024.504022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.665s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.506020] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.506952] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.782s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.507147] env[70013]: DEBUG oslo_concurrency.lockutils [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.507307] env[70013]: INFO nova.compute.manager [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: ef64a05a-b514-4c35-81d3-664ae1ad3ff1] Successfully reverted task state from None on failure for instance. [ 1024.509754] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.871s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.511211] env[70013]: INFO nova.compute.claims [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server [None req-f193c8e9-909e-4c25-bb07-0f41efb74eb4 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-71e7993e-732d-49b9-9e07-f5733cc11272"}]} [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7427, in shelve_instance [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server do_shelve_instance() [ 1024.518078] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7425, in do_shelve_instance [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self._shelve_instance(context, instance, image_id, clean_shutdown, [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7485, in _shelve_instance [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self._shelve_offload_instance( [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7567, in _shelve_offload_instance [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server [None req-947ee1d2-e696-4882-be53-9e6dea76a214 tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ca0d279d-62f0-455b-8f2c-a6644c055a3c"}]} [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1024.520279] env[70013]: ERROR oslo_messaging.rpc.server [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1024.522809] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1024.524559] env[70013]: ERROR oslo_messaging.rpc.server [ 1024.524559] env[70013]: DEBUG oslo_concurrency.lockutils [req-4e2f74c6-387e-4b86-a433-8d4ea5e883a0 req-45cd150d-9efe-429f-82df-6bfa1b09517e service nova] Releasing lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.527040] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231245, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528324} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.527948] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1024.527948] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.527948] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa057661-77b5-4568-a7dc-10b0c344efc5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.535886] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1024.535886] env[70013]: value = "task-4231246" [ 1024.535886] env[70013]: _type = "Task" [ 1024.535886] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.560349] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52aa885d-76f2-fc89-dcd5-9b8120ed43e0, 'name': SearchDatastore_Task, 'duration_secs': 0.016834} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.564999] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.565307] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1024.565671] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231246, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.565816] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1b4a5f3-1612-4380-8116-8c7bdd823ac6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.575652] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1024.575652] env[70013]: value = "task-4231247" [ 1024.575652] env[70013]: _type = "Task" [ 1024.575652] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.590787] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231247, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.805210] env[70013]: DEBUG oslo_concurrency.lockutils [None req-67ddc384-4490-431a-9598-3f2f62200508 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.098s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.023654] env[70013]: DEBUG nova.compute.utils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1025.030791] env[70013]: DEBUG nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1025.030791] env[70013]: DEBUG nova.network.neutron [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1025.047919] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Successfully updated port: 96be8274-aa70-44e5-8a0c-84ad7d0f9737 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.054550] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231246, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07933} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.055061] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.055986] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8badba20-263b-4dfb-ac0d-003a931b1119 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.092327] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.092327] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39da2bb1-0364-4c41-81b5-feee87c17c09 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "30fcb8c7-0947-426e-9aa5-081446e31c28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.625s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.097381] env[70013]: DEBUG nova.policy [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfa9f992483942f689f616b60a27b00d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0896b76438d9456fb1b9fa2f39d0decc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1025.102911] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34779980-5c6b-41c7-b0cf-401ead0bd9f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.128636] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231247, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.130477] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1025.130477] env[70013]: value = "task-4231248" [ 1025.130477] env[70013]: _type = "Task" [ 1025.130477] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.141757] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231248, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.235794] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.236116] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.236342] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.236617] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.236856] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.239532] env[70013]: INFO nova.compute.manager [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Terminating instance [ 1025.585359] env[70013]: DEBUG nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1025.593875] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.594726] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1025.594958] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.612599] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231247, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594738} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.612599] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1025.612599] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.612599] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-520fc8f1-55a5-4496-bd6a-eb87546a6bf4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.623333] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1025.623333] env[70013]: value = "task-4231249" [ 1025.623333] env[70013]: _type = "Task" [ 1025.623333] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.634635] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231249, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.652864] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231248, 'name': ReconfigVM_Task, 'duration_secs': 0.320111} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.652864] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.653365] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e95c090-d854-458b-bf56-72f2d062486a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.665141] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1025.665141] env[70013]: value = "task-4231250" [ 1025.665141] env[70013]: _type = "Task" [ 1025.665141] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.677568] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231250, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.728610] env[70013]: DEBUG nova.compute.manager [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received event network-vif-plugged-96be8274-aa70-44e5-8a0c-84ad7d0f9737 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1025.728985] env[70013]: DEBUG oslo_concurrency.lockutils [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] Acquiring lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.732790] env[70013]: DEBUG oslo_concurrency.lockutils [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.732790] env[70013]: DEBUG oslo_concurrency.lockutils [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.732790] env[70013]: DEBUG nova.compute.manager [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] No waiting events found dispatching network-vif-plugged-96be8274-aa70-44e5-8a0c-84ad7d0f9737 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1025.732790] env[70013]: WARNING nova.compute.manager [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received unexpected event network-vif-plugged-96be8274-aa70-44e5-8a0c-84ad7d0f9737 for instance with vm_state building and task_state spawning. [ 1025.732790] env[70013]: DEBUG nova.compute.manager [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received event network-changed-96be8274-aa70-44e5-8a0c-84ad7d0f9737 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1025.732790] env[70013]: DEBUG nova.compute.manager [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Refreshing instance network info cache due to event network-changed-96be8274-aa70-44e5-8a0c-84ad7d0f9737. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1025.732790] env[70013]: DEBUG oslo_concurrency.lockutils [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] Acquiring lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.749897] env[70013]: DEBUG nova.compute.manager [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1025.750274] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.754091] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2a43c6-8b13-4749-a4df-3189f1a09f41 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.766617] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.769644] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b4a326b-2037-44d2-8bd0-a54a7cae2502 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.781030] env[70013]: DEBUG oslo_vmware.api [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1025.781030] env[70013]: value = "task-4231251" [ 1025.781030] env[70013]: _type = "Task" [ 1025.781030] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.784137] env[70013]: DEBUG nova.network.neutron [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Successfully created port: b3de0526-9f74-4f2b-9519-cdb229f128dd {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.796458] env[70013]: DEBUG oslo_vmware.api [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.133450] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231249, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083081} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.133981] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.134971] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a8a3e9-39c8-4324-ad2f-2345914470d2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.162452] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.163898] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1026.168016] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9eee377-e930-4509-bf3d-c09c2690f6d8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.184339] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dd38ac-d0ea-4b65-91b9-031b0c0a8b5a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.200574] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52ca938-59cb-48a8-b249-a521bc1ed585 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.205112] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1026.205112] env[70013]: value = "task-4231252" [ 1026.205112] env[70013]: _type = "Task" [ 1026.205112] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.205112] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231250, 'name': Rename_Task, 'duration_secs': 0.182895} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.205587] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.208955] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0879a0f4-847e-4a20-ab19-3dc672280a7d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.243309] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d29ee99-078e-4076-8f0f-fec0e40771f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.251800] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231252, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.253751] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1026.253751] env[70013]: value = "task-4231253" [ 1026.253751] env[70013]: _type = "Task" [ 1026.253751] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.260985] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd913db1-a579-439a-a53e-e5fa4e8b0da5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.268452] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231253, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.280541] env[70013]: DEBUG nova.compute.provider_tree [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.294519] env[70013]: DEBUG oslo_vmware.api [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231251, 'name': PowerOffVM_Task, 'duration_secs': 0.227604} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.294815] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1026.295035] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1026.295324] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1153eb80-a12b-4be0-acfa-0db7b661e2e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.369206] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.369535] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.369878] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleting the datastore file [datastore1] 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.370531] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1dc8ca3-3310-4aa2-a1d0-1ebbc575cebe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.378970] env[70013]: DEBUG oslo_vmware.api [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1026.378970] env[70013]: value = "task-4231255" [ 1026.378970] env[70013]: _type = "Task" [ 1026.378970] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.389281] env[70013]: DEBUG oslo_vmware.api [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.548186] env[70013]: DEBUG nova.network.neutron [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updating instance_info_cache with network_info: [{"id": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "address": "fa:16:3e:13:06:6f", "network": {"id": "65f8a088-687e-4cc7-abc0-be598d38a63e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1928781822", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.92", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebcb7b3a-3c", "ovs_interfaceid": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "96be8274-aa70-44e5-8a0c-84ad7d0f9737", "address": "fa:16:3e:7e:be:79", "network": {"id": "72eddfb2-f956-47cc-8bf6-3ca5b03c852b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523543456", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96be8274-aa", "ovs_interfaceid": "96be8274-aa70-44e5-8a0c-84ad7d0f9737", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.555284] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.555582] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.608490] env[70013]: DEBUG nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1026.639254] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1026.639644] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.639987] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1026.640442] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.640784] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1026.641084] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1026.641366] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1026.641628] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1026.641978] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1026.642271] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1026.642589] env[70013]: DEBUG nova.virt.hardware [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1026.643808] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a056ec99-3587-4585-944b-b03066b8971a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.653786] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9795a4d5-bba6-4106-9c81-eec4b3da6781 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.716694] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231252, 'name': ReconfigVM_Task, 'duration_secs': 0.314524} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.717021] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9/0c00737c-3fd4-4ec6-8b47-1ee64096e8f9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.717676] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12de044f-9301-456e-abf6-348aaae05b19 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.727562] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1026.727562] env[70013]: value = "task-4231256" [ 1026.727562] env[70013]: _type = "Task" [ 1026.727562] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.737873] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231256, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.763770] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231253, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.808319] env[70013]: ERROR nova.scheduler.client.report [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [req-91be092e-1880-42c3-9aad-a57ecaf0e8c3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-91be092e-1880-42c3-9aad-a57ecaf0e8c3"}]} [ 1026.808701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.809387] env[70013]: ERROR nova.compute.manager [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Traceback (most recent call last): [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] yield [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] self.set_inventory_for_provider( [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-91be092e-1880-42c3-9aad-a57ecaf0e8c3"}]} [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] During handling of the above exception, another exception occurred: [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Traceback (most recent call last): [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] with self.rt.instance_claim(context, instance, node, allocs, [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] return f(*args, **kwargs) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] self._update(elevated, cn) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] self._update_to_placement(context, compute_node, startup) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] return attempt.get(self._wrap_exception) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] raise value [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] self.reportclient.update_from_provider_tree( [ 1026.809387] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1026.810458] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] with catch_all(pd.uuid): [ 1026.810458] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1026.810458] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] self.gen.throw(typ, value, traceback) [ 1026.810458] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1026.810458] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] raise exception.ResourceProviderSyncFailed() [ 1026.810458] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1026.810458] env[70013]: ERROR nova.compute.manager [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] [ 1026.810458] env[70013]: DEBUG nova.compute.utils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1026.811504] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.016s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.811740] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.811951] env[70013]: INFO nova.compute.manager [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] [instance: 39557c50-866a-4510-b840-b1a6a3e3890e] Successfully reverted task state from None on failure for instance. [ 1026.815302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.518s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.815630] env[70013]: DEBUG nova.objects.instance [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lazy-loading 'resources' on Instance uuid 789fffd9-6725-4bf6-9144-dd603b0a521f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.820034] env[70013]: DEBUG nova.compute.manager [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Build of instance 1ff7f11d-360f-44dd-9ca7-f76c1db905fd was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1026.820492] env[70013]: DEBUG nova.compute.manager [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1026.820802] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "refresh_cache-1ff7f11d-360f-44dd-9ca7-f76c1db905fd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.820997] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "refresh_cache-1ff7f11d-360f-44dd-9ca7-f76c1db905fd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.821231] env[70013]: DEBUG nova.network.neutron [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server [None req-4938dd96-a930-47aa-8a53-bd2354367581 tempest-ServerRescueTestJSONUnderV235-1723903125 tempest-ServerRescueTestJSONUnderV235-1723903125-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9855243a-70dd-434c-846b-338e0d72c52a"}]} [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1026.823824] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1026.825311] env[70013]: ERROR oslo_messaging.rpc.server [ 1026.890460] env[70013]: DEBUG oslo_vmware.api [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157876} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.890771] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.891023] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.891232] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.891414] env[70013]: INFO nova.compute.manager [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1026.891662] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1026.891867] env[70013]: DEBUG nova.compute.manager [-] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1026.891991] env[70013]: DEBUG nova.network.neutron [-] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1027.051074] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Releasing lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.051607] env[70013]: DEBUG nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance network_info: |[{"id": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "address": "fa:16:3e:13:06:6f", "network": {"id": "65f8a088-687e-4cc7-abc0-be598d38a63e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1928781822", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.92", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebcb7b3a-3c", "ovs_interfaceid": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "96be8274-aa70-44e5-8a0c-84ad7d0f9737", "address": "fa:16:3e:7e:be:79", "network": {"id": "72eddfb2-f956-47cc-8bf6-3ca5b03c852b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523543456", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96be8274-aa", "ovs_interfaceid": "96be8274-aa70-44e5-8a0c-84ad7d0f9737", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1027.051974] env[70013]: DEBUG oslo_concurrency.lockutils [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] Acquired lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1027.052191] env[70013]: DEBUG nova.network.neutron [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Refreshing network info cache for port 96be8274-aa70-44e5-8a0c-84ad7d0f9737 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1027.053720] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:06:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebcb7b3a-3c04-471a-b585-af59dfa574f3', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:be:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96be8274-aa70-44e5-8a0c-84ad7d0f9737', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.063735] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1027.064348] env[70013]: DEBUG nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1027.066408] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1027.067588] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38e858ea-7332-4914-bd79-730e47d9fe09 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.094905] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.094905] env[70013]: value = "task-4231257" [ 1027.094905] env[70013]: _type = "Task" [ 1027.094905] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.101324] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231257, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.238918] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231256, 'name': Rename_Task, 'duration_secs': 0.164319} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.239250] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1027.239507] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcde1ca4-2db7-4bcf-ab53-7663f7975b55 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.249956] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1027.249956] env[70013]: value = "task-4231258" [ 1027.249956] env[70013]: _type = "Task" [ 1027.249956] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.266486] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231258, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.270778] env[70013]: DEBUG oslo_vmware.api [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231253, 'name': PowerOnVM_Task, 'duration_secs': 0.531688} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.271431] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1027.271816] env[70013]: INFO nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Took 7.93 seconds to spawn the instance on the hypervisor. [ 1027.272303] env[70013]: DEBUG nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1027.273419] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf8b752-1861-4faf-8851-6c38d211044a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.354251] env[70013]: DEBUG nova.network.neutron [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.357252] env[70013]: DEBUG nova.scheduler.client.report [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1027.375788] env[70013]: DEBUG nova.scheduler.client.report [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1027.376072] env[70013]: DEBUG nova.compute.provider_tree [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1027.390602] env[70013]: DEBUG nova.scheduler.client.report [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1027.417724] env[70013]: DEBUG nova.scheduler.client.report [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1027.604451] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231257, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.609021] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.736305] env[70013]: DEBUG nova.network.neutron [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Successfully updated port: b3de0526-9f74-4f2b-9519-cdb229f128dd {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.766735] env[70013]: DEBUG oslo_vmware.api [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231258, 'name': PowerOnVM_Task, 'duration_secs': 0.45193} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.767335] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1027.767670] env[70013]: DEBUG nova.compute.manager [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1027.771223] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee19c0c2-7f34-47dd-910b-387e1d486f62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.794496] env[70013]: DEBUG nova.network.neutron [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.800404] env[70013]: INFO nova.compute.manager [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Took 34.83 seconds to build instance. [ 1027.859561] env[70013]: DEBUG nova.network.neutron [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updated VIF entry in instance network info cache for port 96be8274-aa70-44e5-8a0c-84ad7d0f9737. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1027.860125] env[70013]: DEBUG nova.network.neutron [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updating instance_info_cache with network_info: [{"id": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "address": "fa:16:3e:13:06:6f", "network": {"id": "65f8a088-687e-4cc7-abc0-be598d38a63e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1928781822", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.92", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebcb7b3a-3c", "ovs_interfaceid": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "96be8274-aa70-44e5-8a0c-84ad7d0f9737", "address": "fa:16:3e:7e:be:79", "network": {"id": "72eddfb2-f956-47cc-8bf6-3ca5b03c852b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-523543456", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96be8274-aa", "ovs_interfaceid": "96be8274-aa70-44e5-8a0c-84ad7d0f9737", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.868871] env[70013]: DEBUG nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Received event network-vif-deleted-4bbb7861-dce4-4ba4-9a50-78e66aa137ef {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1027.869030] env[70013]: INFO nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Neutron deleted interface 4bbb7861-dce4-4ba4-9a50-78e66aa137ef; detaching it from the instance and deleting it from the info cache [ 1027.869175] env[70013]: DEBUG nova.network.neutron [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.910218] env[70013]: DEBUG nova.network.neutron [-] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.941111] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d16372-9eab-4d81-bdc1-f61f224d1614 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.949176] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589b2f9a-7341-41e2-96a9-61f0f5abbee6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.992082] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1e27e3-3869-4820-9bdf-5f862520958a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.000813] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749183e2-c006-4f46-ab2c-25e7378d9b89 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.020437] env[70013]: DEBUG nova.compute.provider_tree [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1028.105934] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231257, 'name': CreateVM_Task, 'duration_secs': 0.538306} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.106154] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.107035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.107243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.107599] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1028.107907] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe6e03a2-5a01-420e-9c44-dea24b066d77 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.113535] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1028.113535] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527b21c4-5ea0-936c-e80c-89b288404bb5" [ 1028.113535] env[70013]: _type = "Task" [ 1028.113535] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.122303] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527b21c4-5ea0-936c-e80c-89b288404bb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.241411] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.241501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.241688] env[70013]: DEBUG nova.network.neutron [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.289684] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.298478] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "refresh_cache-1ff7f11d-360f-44dd-9ca7-f76c1db905fd" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.298702] env[70013]: DEBUG nova.compute.manager [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1028.298891] env[70013]: DEBUG nova.compute.manager [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1028.299142] env[70013]: DEBUG nova.network.neutron [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1028.305018] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9e284abd-6e6c-449c-b2bf-0bc05a1b2d67 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.459s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.317589] env[70013]: DEBUG nova.network.neutron [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1028.363066] env[70013]: DEBUG oslo_concurrency.lockutils [req-a295d447-d9a9-47d1-acbd-cc4a05e5816e req-cf556e9e-6178-402f-ae7a-718899ef5ad0 service nova] Releasing lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.373720] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f066b5f1-6f45-4de1-b27b-b8e0c8726b0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.383484] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4588f1-09e7-425f-b58c-3468ad0124da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.422341] env[70013]: INFO nova.compute.manager [-] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Took 1.53 seconds to deallocate network for instance. [ 1028.422675] env[70013]: DEBUG nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Detach interface failed, port_id=4bbb7861-dce4-4ba4-9a50-78e66aa137ef, reason: Instance 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1028.422968] env[70013]: DEBUG nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Received event network-vif-plugged-b3de0526-9f74-4f2b-9519-cdb229f128dd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1028.423259] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] Acquiring lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.423565] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1028.423803] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.424051] env[70013]: DEBUG nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] No waiting events found dispatching network-vif-plugged-b3de0526-9f74-4f2b-9519-cdb229f128dd {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1028.424298] env[70013]: WARNING nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Received unexpected event network-vif-plugged-b3de0526-9f74-4f2b-9519-cdb229f128dd for instance with vm_state building and task_state spawning. [ 1028.424530] env[70013]: DEBUG nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Received event network-changed-b3de0526-9f74-4f2b-9519-cdb229f128dd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1028.424753] env[70013]: DEBUG nova.compute.manager [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Refreshing instance network info cache due to event network-changed-b3de0526-9f74-4f2b-9519-cdb229f128dd. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1028.424999] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] Acquiring lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.542552] env[70013]: ERROR nova.scheduler.client.report [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [req-9cf38602-1b07-4c78-b3ce-ac9110fd60cb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9cf38602-1b07-4c78-b3ce-ac9110fd60cb"}]} [ 1028.542934] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.728s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.543605] env[70013]: ERROR nova.compute.manager [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Traceback (most recent call last): [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] yield [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self.set_inventory_for_provider( [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9cf38602-1b07-4c78-b3ce-ac9110fd60cb"}]} [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] During handling of the above exception, another exception occurred: [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Traceback (most recent call last): [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self._delete_instance(context, instance, bdms) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self._complete_deletion(context, instance) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self._update_resource_tracker(context, instance) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self.rt.update_usage(context, instance, instance.node) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] return f(*args, **kwargs) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self._update_to_placement(context, compute_node, startup) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] return attempt.get(self._wrap_exception) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1028.543605] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] raise value [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self.reportclient.update_from_provider_tree( [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] with catch_all(pd.uuid): [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] self.gen.throw(typ, value, traceback) [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] raise exception.ResourceProviderSyncFailed() [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1028.544684] env[70013]: ERROR nova.compute.manager [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] [ 1028.546385] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.146s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1028.546577] env[70013]: DEBUG oslo_concurrency.lockutils [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.546745] env[70013]: INFO nova.compute.manager [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Successfully reverted task state from None on failure for instance. [ 1028.549053] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.607s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1028.550591] env[70013]: INFO nova.compute.claims [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server [None req-62a18c5d-b81c-43d1-ba1a-6c72fe4c5131 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e35264e5-8c42-4c01-a1ec-ad8e2017dcb2"}]} [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1028.553824] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1028.555500] env[70013]: ERROR oslo_messaging.rpc.server [ 1028.625805] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527b21c4-5ea0-936c-e80c-89b288404bb5, 'name': SearchDatastore_Task, 'duration_secs': 0.039651} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.626205] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.626491] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1028.626652] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.626742] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.627104] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1028.627321] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58085fec-1d20-47cf-a8a6-c65d527caa7c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.637089] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1028.637286] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1028.638039] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4c20bab-e4ef-4925-bd55-dcc9f107bef4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.643505] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1028.643505] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52df8615-300b-74e2-c772-0623884b3953" [ 1028.643505] env[70013]: _type = "Task" [ 1028.643505] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.654520] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52df8615-300b-74e2-c772-0623884b3953, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.784444] env[70013]: DEBUG nova.network.neutron [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1028.820965] env[70013]: DEBUG nova.network.neutron [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.933301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.951135] env[70013]: DEBUG nova.network.neutron [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Updating instance_info_cache with network_info: [{"id": "b3de0526-9f74-4f2b-9519-cdb229f128dd", "address": "fa:16:3e:9d:3e:d1", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3de0526-9f", "ovs_interfaceid": "b3de0526-9f74-4f2b-9519-cdb229f128dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.056168] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.191s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.156034] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52df8615-300b-74e2-c772-0623884b3953, 'name': SearchDatastore_Task, 'duration_secs': 0.018281} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.156793] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f982dc6a-c929-40b0-92d8-e07eba33e17e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.163244] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1029.163244] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e318b7-a97c-3143-603e-967e514f2209" [ 1029.163244] env[70013]: _type = "Task" [ 1029.163244] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.172301] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e318b7-a97c-3143-603e-967e514f2209, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.324028] env[70013]: INFO nova.compute.manager [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 1ff7f11d-360f-44dd-9ca7-f76c1db905fd] Took 1.02 seconds to deallocate network for instance. [ 1029.407139] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.407792] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.407887] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.408144] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.408359] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.410740] env[70013]: INFO nova.compute.manager [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Terminating instance [ 1029.454180] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.454569] env[70013]: DEBUG nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance network_info: |[{"id": "b3de0526-9f74-4f2b-9519-cdb229f128dd", "address": "fa:16:3e:9d:3e:d1", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3de0526-9f", "ovs_interfaceid": "b3de0526-9f74-4f2b-9519-cdb229f128dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1029.454965] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] Acquired lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.455183] env[70013]: DEBUG nova.network.neutron [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Refreshing network info cache for port b3de0526-9f74-4f2b-9519-cdb229f128dd {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.456466] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:3e:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac7039c0-3374-4c08-87fc-af2449b48b02', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3de0526-9f74-4f2b-9519-cdb229f128dd', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.464569] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1029.468186] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1029.468664] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-614a09d2-3120-4bbf-a088-b7fc77187ca2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.492660] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.492660] env[70013]: value = "task-4231259" [ 1029.492660] env[70013]: _type = "Task" [ 1029.492660] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.502871] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231259, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.583529] env[70013]: DEBUG nova.scheduler.client.report [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1029.599519] env[70013]: DEBUG nova.scheduler.client.report [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1029.599790] env[70013]: DEBUG nova.compute.provider_tree [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1029.619408] env[70013]: DEBUG nova.scheduler.client.report [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1029.641845] env[70013]: DEBUG nova.scheduler.client.report [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1029.679068] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e318b7-a97c-3143-603e-967e514f2209, 'name': SearchDatastore_Task, 'duration_secs': 0.036304} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.679636] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.679636] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 181238eb-cb0c-4740-9896-ca745b53ebe8/181238eb-cb0c-4740-9896-ca745b53ebe8.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1029.679882] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70a554ed-827c-45c1-9d76-dbea7ca76316 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.690559] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1029.690559] env[70013]: value = "task-4231260" [ 1029.690559] env[70013]: _type = "Task" [ 1029.690559] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.699741] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.833703] env[70013]: DEBUG nova.network.neutron [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Updated VIF entry in instance network info cache for port b3de0526-9f74-4f2b-9519-cdb229f128dd. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.834219] env[70013]: DEBUG nova.network.neutron [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Updating instance_info_cache with network_info: [{"id": "b3de0526-9f74-4f2b-9519-cdb229f128dd", "address": "fa:16:3e:9d:3e:d1", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3de0526-9f", "ovs_interfaceid": "b3de0526-9f74-4f2b-9519-cdb229f128dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.898358] env[70013]: DEBUG nova.compute.manager [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Received event network-changed-3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1029.898555] env[70013]: DEBUG nova.compute.manager [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Refreshing instance network info cache due to event network-changed-3d1c82dc-c975-4156-8d68-368f68ec3f87. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1029.898780] env[70013]: DEBUG oslo_concurrency.lockutils [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] Acquiring lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.898927] env[70013]: DEBUG oslo_concurrency.lockutils [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] Acquired lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.899164] env[70013]: DEBUG nova.network.neutron [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Refreshing network info cache for port 3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.918262] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "refresh_cache-0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.918262] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquired lock "refresh_cache-0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.918262] env[70013]: DEBUG nova.network.neutron [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1030.006381] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231259, 'name': CreateVM_Task, 'duration_secs': 0.363545} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.006563] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.007317] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.007488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.007815] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1030.008090] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f9c8234-bbe3-4a89-9bbd-be172a4205bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.016294] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1030.016294] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52094d1e-41ac-7fb1-ccbe-78737cc423fb" [ 1030.016294] env[70013]: _type = "Task" [ 1030.016294] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.026730] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52094d1e-41ac-7fb1-ccbe-78737cc423fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.120016] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cbc2e8-db72-4b8e-9b60-f5d19e00b138 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.131619] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d915e3-a0cf-44d3-947c-87f9b63e546a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.170459] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d97169-c8de-4040-a13d-c5ba4a332ec5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.180759] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26bdc81-ec4c-4147-80aa-3aa8ed5ae374 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.197863] env[70013]: DEBUG nova.compute.provider_tree [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.209749] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231260, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.341265] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6de29f9-ceca-41c6-aecc-d093c8aa4fc9 req-ed5c8340-5413-4444-9d0b-f3caa664e868 service nova] Releasing lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.364126] env[70013]: INFO nova.scheduler.client.report [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted allocations for instance 1ff7f11d-360f-44dd-9ca7-f76c1db905fd [ 1030.453501] env[70013]: DEBUG nova.network.neutron [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1030.524575] env[70013]: DEBUG nova.network.neutron [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.530500] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52094d1e-41ac-7fb1-ccbe-78737cc423fb, 'name': SearchDatastore_Task, 'duration_secs': 0.016495} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.531256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.531529] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.531819] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.532075] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.532317] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.532663] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b20c559-9c6f-4f77-b10f-70dcc4662eaa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.543140] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.543373] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1030.544174] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01404a9a-8bcd-45ad-b4c7-67b3210f4463 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.555540] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1030.555540] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5270c838-6dd7-de0f-d667-d4d5746149b8" [ 1030.555540] env[70013]: _type = "Task" [ 1030.555540] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.570413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.570829] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5270c838-6dd7-de0f-d667-d4d5746149b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.714348] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.796557} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.714687] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 181238eb-cb0c-4740-9896-ca745b53ebe8/181238eb-cb0c-4740-9896-ca745b53ebe8.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1030.714839] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1030.715119] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-419fc7ce-1f1f-4126-b869-13ac70508c5a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.724103] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1030.724103] env[70013]: value = "task-4231261" [ 1030.724103] env[70013]: _type = "Task" [ 1030.724103] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.734775] env[70013]: ERROR nova.scheduler.client.report [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [req-ed7b2fa8-07ab-4d82-9835-1c1e68bbbefc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ed7b2fa8-07ab-4d82-9835-1c1e68bbbefc"}]} [ 1030.735092] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.186s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.735695] env[70013]: ERROR nova.compute.manager [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Traceback (most recent call last): [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] yield [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] self.set_inventory_for_provider( [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ed7b2fa8-07ab-4d82-9835-1c1e68bbbefc"}]} [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] During handling of the above exception, another exception occurred: [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Traceback (most recent call last): [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] with self.rt.instance_claim(context, instance, node, allocs, [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] return f(*args, **kwargs) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] self._update(elevated, cn) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] self._update_to_placement(context, compute_node, startup) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] return attempt.get(self._wrap_exception) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] raise value [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] self.reportclient.update_from_provider_tree( [ 1030.735695] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1030.738761] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] with catch_all(pd.uuid): [ 1030.738761] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1030.738761] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] self.gen.throw(typ, value, traceback) [ 1030.738761] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1030.738761] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] raise exception.ResourceProviderSyncFailed() [ 1030.738761] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1030.738761] env[70013]: ERROR nova.compute.manager [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] [ 1030.738761] env[70013]: DEBUG nova.compute.utils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1030.740672] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.847s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.740672] env[70013]: DEBUG nova.objects.instance [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lazy-loading 'resources' on Instance uuid 9a61dedd-3764-4bd9-a300-480cc7d14a21 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.745354] env[70013]: DEBUG nova.compute.manager [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Build of instance 74f39e88-b034-4e45-ae1c-453d644c093b was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1030.745851] env[70013]: DEBUG nova.compute.manager [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1030.746129] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-74f39e88-b034-4e45-ae1c-453d644c093b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.746318] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-74f39e88-b034-4e45-ae1c-453d644c093b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.746511] env[70013]: DEBUG nova.network.neutron [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1030.749049] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.771628] env[70013]: DEBUG nova.scheduler.client.report [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1030.787620] env[70013]: DEBUG nova.scheduler.client.report [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1030.787909] env[70013]: DEBUG nova.compute.provider_tree [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.807039] env[70013]: DEBUG nova.scheduler.client.report [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1030.809908] env[70013]: DEBUG nova.network.neutron [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Updated VIF entry in instance network info cache for port 3d1c82dc-c975-4156-8d68-368f68ec3f87. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1030.810530] env[70013]: DEBUG nova.network.neutron [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Updating instance_info_cache with network_info: [{"id": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "address": "fa:16:3e:d3:e4:4b", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d1c82dc-c9", "ovs_interfaceid": "3d1c82dc-c975-4156-8d68-368f68ec3f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.824726] env[70013]: DEBUG nova.scheduler.client.report [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1030.875449] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d885f5c2-d942-46a1-859b-46bad5e4929a tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "1ff7f11d-360f-44dd-9ca7-f76c1db905fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.264s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.035359] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Releasing lock "refresh_cache-0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.036031] env[70013]: DEBUG nova.compute.manager [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1031.036299] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.038215] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac581e9c-dd1e-4894-a1ff-b93ac667bc69 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.050219] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.052731] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f1da7a9-2bfb-45de-94b9-50c28301fc4c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.061992] env[70013]: DEBUG oslo_vmware.api [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1031.061992] env[70013]: value = "task-4231262" [ 1031.061992] env[70013]: _type = "Task" [ 1031.061992] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.074016] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5270c838-6dd7-de0f-d667-d4d5746149b8, 'name': SearchDatastore_Task, 'duration_secs': 0.014391} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.075522] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c4a5bb2-c994-4cbf-9a42-615227a51288 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.081258] env[70013]: DEBUG oslo_vmware.api [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231262, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.085120] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1031.085120] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5204bc12-898f-88e4-9a0a-9c44bb224cd1" [ 1031.085120] env[70013]: _type = "Task" [ 1031.085120] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.098207] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5204bc12-898f-88e4-9a0a-9c44bb224cd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.236742] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077333} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.237185] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.238141] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9492589c-a8e4-4aa9-9e0b-b8d34c47fbf9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.270115] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 181238eb-cb0c-4740-9896-ca745b53ebe8/181238eb-cb0c-4740-9896-ca745b53ebe8.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.271074] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1edfb8dd-bcd9-4e80-9af6-b44ec753c68c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.294880] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1031.294880] env[70013]: value = "task-4231263" [ 1031.294880] env[70013]: _type = "Task" [ 1031.294880] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.309553] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231263, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.315878] env[70013]: DEBUG oslo_concurrency.lockutils [req-4725b681-dc01-4c5f-8965-793b616734ff req-075ba86b-5c43-4f99-af5e-cf2a2da03aa7 service nova] Releasing lock "refresh_cache-1230ca35-1512-4464-85fc-d3b4ab05eac1" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.318524] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a554ed6-fc2d-4d6b-a01c-a2365f923d42 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.322824] env[70013]: DEBUG nova.network.neutron [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1031.331528] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03ac8a5-292d-4ad3-a28c-2ce9a65af2e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.368991] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255c4147-82e2-470e-a29d-64670a137286 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.377676] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a8a413-381e-4bde-8a45-10b8b7df4489 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.395439] env[70013]: DEBUG nova.compute.provider_tree [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1031.426485] env[70013]: DEBUG nova.network.neutron [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.579985] env[70013]: DEBUG oslo_vmware.api [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231262, 'name': PowerOffVM_Task, 'duration_secs': 0.130003} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.580463] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.580745] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1031.581151] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b01bb589-4e42-45f2-b1f8-bce3bdda69a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.604649] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5204bc12-898f-88e4-9a0a-9c44bb224cd1, 'name': SearchDatastore_Task, 'duration_secs': 0.01415} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.604951] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.605509] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9026841c-7bda-41e5-a4ac-03d0d3e37560/9026841c-7bda-41e5-a4ac-03d0d3e37560.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1031.605868] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1820a2d0-b9da-4a54-a31e-440a16748d51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.610222] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1031.610569] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1031.611117] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Deleting the datastore file [datastore1] 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1031.611739] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a314f69-5c91-483e-8ff9-2d35e65e9999 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.617464] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1031.617464] env[70013]: value = "task-4231265" [ 1031.617464] env[70013]: _type = "Task" [ 1031.617464] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.624850] env[70013]: DEBUG oslo_vmware.api [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for the task: (returnval){ [ 1031.624850] env[70013]: value = "task-4231266" [ 1031.624850] env[70013]: _type = "Task" [ 1031.624850] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.633700] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231265, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.641344] env[70013]: DEBUG oslo_vmware.api [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.806263] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231263, 'name': ReconfigVM_Task, 'duration_secs': 0.388552} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.806666] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 181238eb-cb0c-4740-9896-ca745b53ebe8/181238eb-cb0c-4740-9896-ca745b53ebe8.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.807407] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36fe535d-00bf-4ef0-a958-36347baa5683 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.814795] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1031.814795] env[70013]: value = "task-4231267" [ 1031.814795] env[70013]: _type = "Task" [ 1031.814795] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.824721] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231267, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.922497] env[70013]: ERROR nova.scheduler.client.report [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [req-50ab243d-fc99-4583-8f01-ac2bfba78d9c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-50ab243d-fc99-4583-8f01-ac2bfba78d9c"}]} [ 1031.922922] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.183s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.923567] env[70013]: ERROR nova.compute.manager [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Traceback (most recent call last): [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] yield [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self.set_inventory_for_provider( [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-50ab243d-fc99-4583-8f01-ac2bfba78d9c"}]} [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] During handling of the above exception, another exception occurred: [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Traceback (most recent call last): [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self._delete_instance(context, instance, bdms) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self._complete_deletion(context, instance) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self._update_resource_tracker(context, instance) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self.rt.update_usage(context, instance, instance.node) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] return f(*args, **kwargs) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self._update_to_placement(context, compute_node, startup) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] return attempt.get(self._wrap_exception) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1031.923567] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] raise value [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self.reportclient.update_from_provider_tree( [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] with catch_all(pd.uuid): [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] self.gen.throw(typ, value, traceback) [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] raise exception.ResourceProviderSyncFailed() [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1031.925193] env[70013]: ERROR nova.compute.manager [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] [ 1031.927707] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.515s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.928364] env[70013]: DEBUG nova.objects.instance [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lazy-loading 'resources' on Instance uuid eab0c393-4d3d-4659-8225-c82ddadb330b {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.932368] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-74f39e88-b034-4e45-ae1c-453d644c093b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.932368] env[70013]: DEBUG nova.compute.manager [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1031.932368] env[70013]: DEBUG nova.compute.manager [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1031.932368] env[70013]: DEBUG nova.network.neutron [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1031.949967] env[70013]: DEBUG nova.network.neutron [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.129441] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231265, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.141061] env[70013]: DEBUG oslo_vmware.api [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Task: {'id': task-4231266, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129393} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.141061] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.141321] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1032.141480] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1032.141680] env[70013]: INFO nova.compute.manager [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1032.141986] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1032.142225] env[70013]: DEBUG nova.compute.manager [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1032.142331] env[70013]: DEBUG nova.network.neutron [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1032.163913] env[70013]: DEBUG nova.network.neutron [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.327206] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231267, 'name': Rename_Task, 'duration_secs': 0.225938} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.327575] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.327827] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-141fb5d3-848f-4065-9ad7-a30efe4b28a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.335840] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1032.335840] env[70013]: value = "task-4231268" [ 1032.335840] env[70013]: _type = "Task" [ 1032.335840] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.345021] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.433516] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.725s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.453296] env[70013]: DEBUG nova.scheduler.client.report [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1032.455443] env[70013]: DEBUG nova.network.neutron [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.468907] env[70013]: DEBUG nova.scheduler.client.report [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1032.469155] env[70013]: DEBUG nova.compute.provider_tree [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1032.480843] env[70013]: DEBUG nova.scheduler.client.report [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1032.498101] env[70013]: DEBUG nova.scheduler.client.report [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1032.630057] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231265, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604567} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.633168] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 9026841c-7bda-41e5-a4ac-03d0d3e37560/9026841c-7bda-41e5-a4ac-03d0d3e37560.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.633449] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.634152] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9344c233-f8f0-4b3e-8d86-eb074b800fe8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.643742] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1032.643742] env[70013]: value = "task-4231269" [ 1032.643742] env[70013]: _type = "Task" [ 1032.643742] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.657535] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231269, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.667423] env[70013]: DEBUG nova.network.neutron [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.851525] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.942017] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c5c90a-e582-4af9-82ae-b862bdad8fb5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.952483] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff24048-683c-4e01-9510-c7ced5a057dc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.957649] env[70013]: INFO nova.compute.manager [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 74f39e88-b034-4e45-ae1c-453d644c093b] Took 1.03 seconds to deallocate network for instance. [ 1033.019277] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4c3a08-ac61-4922-bde3-1935e23b3048 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.029791] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "f0acce5d-d747-43b8-a5c8-ae10c10605ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.029957] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "f0acce5d-d747-43b8-a5c8-ae10c10605ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.040155] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf31457-0c11-43a4-ae6b-2220dbd574d4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.065096] env[70013]: DEBUG nova.compute.provider_tree [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1033.153860] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231269, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.308426} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.154146] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1033.154939] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4fb7d1-ad5e-4bd7-b68e-bdc37b2e2dca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.170089] env[70013]: INFO nova.compute.manager [-] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Took 1.03 seconds to deallocate network for instance. [ 1033.179154] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 9026841c-7bda-41e5-a4ac-03d0d3e37560/9026841c-7bda-41e5-a4ac-03d0d3e37560.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.181066] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f796bf9-9104-4666-8af9-023aba3b2df9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.206875] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1033.206875] env[70013]: value = "task-4231270" [ 1033.206875] env[70013]: _type = "Task" [ 1033.206875] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.221325] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231270, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.279562] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.279828] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.280083] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.280285] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.280457] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.282690] env[70013]: INFO nova.compute.manager [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Terminating instance [ 1033.347041] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.534316] env[70013]: DEBUG nova.compute.manager [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1033.589855] env[70013]: ERROR nova.scheduler.client.report [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [req-087732a9-8a3e-47f4-845e-ba5b401e31e9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-087732a9-8a3e-47f4-845e-ba5b401e31e9"}]} [ 1033.590238] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.663s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.590861] env[70013]: ERROR nova.compute.manager [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Traceback (most recent call last): [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] yield [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self.set_inventory_for_provider( [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-087732a9-8a3e-47f4-845e-ba5b401e31e9"}]} [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] During handling of the above exception, another exception occurred: [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Traceback (most recent call last): [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self._delete_instance(context, instance, bdms) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self._complete_deletion(context, instance) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self._update_resource_tracker(context, instance) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self.rt.update_usage(context, instance, instance.node) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] return f(*args, **kwargs) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self._update_to_placement(context, compute_node, startup) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] return attempt.get(self._wrap_exception) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1033.590861] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] raise value [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self.reportclient.update_from_provider_tree( [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] with catch_all(pd.uuid): [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] self.gen.throw(typ, value, traceback) [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] raise exception.ResourceProviderSyncFailed() [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1033.591955] env[70013]: ERROR nova.compute.manager [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] [ 1033.593334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.089s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.593555] env[70013]: DEBUG nova.objects.instance [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lazy-loading 'resources' on Instance uuid 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.701486] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.721505] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.786241] env[70013]: DEBUG nova.compute.manager [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1033.786404] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1033.787340] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0609ecf-6744-4e06-8fcc-9510b3b452c8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.797656] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.798802] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c1a2acf-2962-42eb-ad1a-3779b3e7a508 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.806500] env[70013]: DEBUG oslo_vmware.api [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 1033.806500] env[70013]: value = "task-4231271" [ 1033.806500] env[70013]: _type = "Task" [ 1033.806500] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.817239] env[70013]: DEBUG oslo_vmware.api [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.848549] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231268, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.944805] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.000610] env[70013]: INFO nova.scheduler.client.report [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted allocations for instance 74f39e88-b034-4e45-ae1c-453d644c093b [ 1034.058663] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.099216] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.388s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.121072] env[70013]: DEBUG nova.scheduler.client.report [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1034.137059] env[70013]: DEBUG nova.scheduler.client.report [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1034.137756] env[70013]: DEBUG nova.compute.provider_tree [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1034.151800] env[70013]: DEBUG nova.scheduler.client.report [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1034.175974] env[70013]: DEBUG nova.scheduler.client.report [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1034.222886] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231270, 'name': ReconfigVM_Task, 'duration_secs': 0.804062} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.225839] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 9026841c-7bda-41e5-a4ac-03d0d3e37560/9026841c-7bda-41e5-a4ac-03d0d3e37560.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.227921] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a07adc1-ec08-4752-b257-60fa9e082022 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.235185] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1034.235185] env[70013]: value = "task-4231272" [ 1034.235185] env[70013]: _type = "Task" [ 1034.235185] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.250765] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231272, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.317924] env[70013]: DEBUG oslo_vmware.api [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231271, 'name': PowerOffVM_Task, 'duration_secs': 0.382864} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.320597] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.320797] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1034.321318] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff15afe5-dd11-4eb2-9f5e-0c2b1210b069 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.349260] env[70013]: DEBUG oslo_vmware.api [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231268, 'name': PowerOnVM_Task, 'duration_secs': 1.7736} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.349541] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.349759] env[70013]: INFO nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Took 12.44 seconds to spawn the instance on the hypervisor. [ 1034.349924] env[70013]: DEBUG nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1034.350779] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee4853a-4168-440b-811d-7fc62b44f724 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.400623] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1034.400824] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1034.401176] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Deleting the datastore file [datastore2] c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.404673] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a606f1a2-3a8f-4a2b-b422-03fcf9639a56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.413307] env[70013]: DEBUG oslo_vmware.api [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 1034.413307] env[70013]: value = "task-4231274" [ 1034.413307] env[70013]: _type = "Task" [ 1034.413307] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.426714] env[70013]: DEBUG oslo_vmware.api [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.512843] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74b2cc19-88d1-4be6-9452-9c8561a5374b tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "74f39e88-b034-4e45-ae1c-453d644c093b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.601s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.693704] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76095d9c-7b25-4ef9-919c-26a38dfb24c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.701742] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6320e81-77e1-4a9c-876b-70fe5a0d9117 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.734124] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17719f20-7ee1-4527-af4c-73fd206dd767 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.749491] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99160f15-8d12-4c8f-80a6-7d803ce620d0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.753797] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231272, 'name': Rename_Task, 'duration_secs': 0.209081} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.754092] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.754843] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a301dc3b-a3c6-407f-bd04-ddf0864efc13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.767997] env[70013]: DEBUG nova.compute.provider_tree [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1034.776276] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1034.776276] env[70013]: value = "task-4231275" [ 1034.776276] env[70013]: _type = "Task" [ 1034.776276] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.787589] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.876336] env[70013]: INFO nova.compute.manager [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Took 37.37 seconds to build instance. [ 1034.926200] env[70013]: DEBUG oslo_vmware.api [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320957} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.926506] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.926735] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.926951] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.927171] env[70013]: INFO nova.compute.manager [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1034.927467] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1034.927799] env[70013]: DEBUG nova.compute.manager [-] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1034.927881] env[70013]: DEBUG nova.network.neutron [-] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1035.287169] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231275, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.289516] env[70013]: DEBUG nova.compute.manager [req-45d0ebec-cf01-4a72-9397-a03be3a6959d req-a8a8b576-ff75-4a8f-bfbd-9410ac4ddc3a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Received event network-vif-deleted-36c36a41-a7ab-45a3-a1c4-be24ecb013e7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1035.289707] env[70013]: INFO nova.compute.manager [req-45d0ebec-cf01-4a72-9397-a03be3a6959d req-a8a8b576-ff75-4a8f-bfbd-9410ac4ddc3a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Neutron deleted interface 36c36a41-a7ab-45a3-a1c4-be24ecb013e7; detaching it from the instance and deleting it from the info cache [ 1035.290023] env[70013]: DEBUG nova.network.neutron [req-45d0ebec-cf01-4a72-9397-a03be3a6959d req-a8a8b576-ff75-4a8f-bfbd-9410ac4ddc3a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.291952] env[70013]: ERROR nova.scheduler.client.report [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [req-f23150ca-7490-47b1-8a51-526f1e183fe9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f23150ca-7490-47b1-8a51-526f1e183fe9"}]} [ 1035.292370] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.292922] env[70013]: ERROR nova.compute.manager [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Traceback (most recent call last): [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] yield [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self.set_inventory_for_provider( [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f23150ca-7490-47b1-8a51-526f1e183fe9"}]} [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] During handling of the above exception, another exception occurred: [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Traceback (most recent call last): [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self._delete_instance(context, instance, bdms) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self._complete_deletion(context, instance) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self._update_resource_tracker(context, instance) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self.rt.update_usage(context, instance, instance.node) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] return f(*args, **kwargs) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self._update_to_placement(context, compute_node, startup) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] return attempt.get(self._wrap_exception) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1035.292922] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] raise value [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self.reportclient.update_from_provider_tree( [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] with catch_all(pd.uuid): [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] self.gen.throw(typ, value, traceback) [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] raise exception.ResourceProviderSyncFailed() [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1035.294132] env[70013]: ERROR nova.compute.manager [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] [ 1035.295247] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.533s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.297263] env[70013]: INFO nova.compute.claims [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.378048] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4b5dd892-da0c-47e7-96ba-be93e4d747b1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.882s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.612589] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.695018] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "8dbef727-13e9-4231-95a8-65d015ee13be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.695390] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.695743] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.696045] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.696283] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.698966] env[70013]: INFO nova.compute.manager [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Terminating instance [ 1035.706915] env[70013]: DEBUG nova.network.neutron [-] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.788455] env[70013]: DEBUG oslo_vmware.api [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231275, 'name': PowerOnVM_Task, 'duration_secs': 0.879751} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.788756] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.788959] env[70013]: INFO nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Took 9.18 seconds to spawn the instance on the hypervisor. [ 1035.789158] env[70013]: DEBUG nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1035.790067] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5287e29e-bb5f-40ea-8d24-39ed28974344 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.801457] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.070s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.805892] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ed3abd0-b752-44f4-b4fc-f3622ad65d5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.816062] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb0cb58-5392-46bc-a6ef-ce45e880e5df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.850348] env[70013]: DEBUG nova.compute.manager [req-45d0ebec-cf01-4a72-9397-a03be3a6959d req-a8a8b576-ff75-4a8f-bfbd-9410ac4ddc3a service nova] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Detach interface failed, port_id=36c36a41-a7ab-45a3-a1c4-be24ecb013e7, reason: Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1035.915753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "181238eb-cb0c-4740-9896-ca745b53ebe8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.916054] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.916300] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.916490] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.916661] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.919076] env[70013]: INFO nova.compute.manager [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Terminating instance [ 1036.204244] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.204244] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1036.204244] env[70013]: DEBUG nova.network.neutron [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.209602] env[70013]: INFO nova.compute.manager [-] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Took 1.28 seconds to deallocate network for instance. [ 1036.318774] env[70013]: INFO nova.compute.manager [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Took 38.12 seconds to build instance. [ 1036.347488] env[70013]: DEBUG nova.scheduler.client.report [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1036.367518] env[70013]: DEBUG nova.scheduler.client.report [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1036.367790] env[70013]: DEBUG nova.compute.provider_tree [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1036.383161] env[70013]: DEBUG nova.scheduler.client.report [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1036.409526] env[70013]: DEBUG nova.scheduler.client.report [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1036.424187] env[70013]: DEBUG nova.compute.manager [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1036.424438] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.425409] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52274f6-dbcb-4d8f-b3aa-fdf45c8ed411 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.435559] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.435559] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fa459be-d8f1-4a47-aea4-8629a8044a46 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.442614] env[70013]: DEBUG oslo_vmware.api [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1036.442614] env[70013]: value = "task-4231276" [ 1036.442614] env[70013]: _type = "Task" [ 1036.442614] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.456126] env[70013]: DEBUG oslo_vmware.api [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231276, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.718455] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.737661] env[70013]: DEBUG nova.network.neutron [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1036.770968] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.825887] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b998c6e0-6519-44dc-96da-793fda77b35f tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.642s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.826665] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.056s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.826894] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.827123] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.827299] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.829548] env[70013]: INFO nova.compute.manager [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Terminating instance [ 1036.952774] env[70013]: DEBUG oslo_vmware.api [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231276, 'name': PowerOffVM_Task, 'duration_secs': 0.276474} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.953069] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.953250] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.953552] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80c74c0a-c5ac-489d-8017-6692d15f1ead {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.975015] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "eab0c393-4d3d-4659-8225-c82ddadb330b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.975278] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.975486] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.975672] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.975835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.978232] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d05dc8-41ae-436f-ab7d-901eb7257c1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.981462] env[70013]: INFO nova.compute.manager [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Terminating instance [ 1036.989637] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f900dc3-051a-43e2-a808-4f405af28e7c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.029929] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e6168a-4060-4df2-b393-2110fc580910 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.042152] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8366db-14bb-4e40-8ec4-2521cebb42c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.060052] env[70013]: DEBUG nova.compute.provider_tree [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1037.062679] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.062934] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.063147] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Deleting the datastore file [datastore1] 181238eb-cb0c-4740-9896-ca745b53ebe8 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.063676] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e7283eb-ee2e-47a7-863d-7aa730ceb27e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.070788] env[70013]: DEBUG oslo_vmware.api [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for the task: (returnval){ [ 1037.070788] env[70013]: value = "task-4231278" [ 1037.070788] env[70013]: _type = "Task" [ 1037.070788] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.081762] env[70013]: DEBUG oslo_vmware.api [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231278, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.087155] env[70013]: DEBUG nova.network.neutron [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.328196] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1037.333212] env[70013]: DEBUG nova.compute.manager [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1037.333269] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.334199] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed29a3d-25ef-4e17-9f0b-ba0a7252fcef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.343752] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.344225] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee72c44c-c66b-468a-8393-014d315ad61d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.352427] env[70013]: DEBUG oslo_vmware.api [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1037.352427] env[70013]: value = "task-4231279" [ 1037.352427] env[70013]: _type = "Task" [ 1037.352427] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.361056] env[70013]: DEBUG oslo_vmware.api [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.496174] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.496658] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.497070] env[70013]: DEBUG nova.network.neutron [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.582564] env[70013]: DEBUG oslo_vmware.api [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Task: {'id': task-4231278, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152269} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.582843] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1037.583048] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1037.583321] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1037.583523] env[70013]: INFO nova.compute.manager [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1037.584152] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1037.584152] env[70013]: DEBUG nova.compute.manager [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1037.584152] env[70013]: DEBUG nova.network.neutron [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1037.589456] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-8dbef727-13e9-4231-95a8-65d015ee13be" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1037.589848] env[70013]: DEBUG nova.compute.manager [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1037.590018] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.590313] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7b8ffd96-2820-48ad-91ad-3d60d79b1155 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.593790] env[70013]: ERROR nova.scheduler.client.report [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [req-cf282be4-782c-4bd8-93bf-d6f81134dcc6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-cf282be4-782c-4bd8-93bf-d6f81134dcc6"}]} [ 1037.594180] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.594809] env[70013]: ERROR nova.compute.manager [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Traceback (most recent call last): [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] yield [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] self.set_inventory_for_provider( [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-cf282be4-782c-4bd8-93bf-d6f81134dcc6"}]} [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] During handling of the above exception, another exception occurred: [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Traceback (most recent call last): [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] with self.rt.instance_claim(context, instance, node, allocs, [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] return f(*args, **kwargs) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] self._update(elevated, cn) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] self._update_to_placement(context, compute_node, startup) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] return attempt.get(self._wrap_exception) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] raise value [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] self.reportclient.update_from_provider_tree( [ 1037.594809] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1037.595727] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] with catch_all(pd.uuid): [ 1037.595727] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1037.595727] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] self.gen.throw(typ, value, traceback) [ 1037.595727] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1037.595727] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] raise exception.ResourceProviderSyncFailed() [ 1037.595727] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1037.595727] env[70013]: ERROR nova.compute.manager [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] [ 1037.595727] env[70013]: DEBUG nova.compute.utils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1037.597032] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.797s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1037.598685] env[70013]: INFO nova.compute.claims [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.601745] env[70013]: DEBUG nova.compute.manager [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Build of instance 5661914f-ccd6-48da-bccc-a6b9065338c3 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1037.602235] env[70013]: DEBUG nova.compute.manager [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1037.602467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "refresh_cache-5661914f-ccd6-48da-bccc-a6b9065338c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.602644] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired lock "refresh_cache-5661914f-ccd6-48da-bccc-a6b9065338c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.602770] env[70013]: DEBUG nova.network.neutron [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.608338] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b83215-6d9c-479b-adb2-6e90cf2da2ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.647084] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8dbef727-13e9-4231-95a8-65d015ee13be could not be found. [ 1037.647544] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1037.648064] env[70013]: INFO nova.compute.manager [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1037.648377] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1037.649184] env[70013]: DEBUG nova.compute.manager [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1037.649428] env[70013]: DEBUG nova.network.neutron [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1037.670090] env[70013]: DEBUG nova.network.neutron [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1037.867908] env[70013]: DEBUG oslo_vmware.api [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231279, 'name': PowerOffVM_Task, 'duration_secs': 0.181192} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.867908] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.867908] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.871066] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5c3a95c-dc07-4b3b-bbce-fbe9370aba1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.939386] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.939643] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.939839] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleting the datastore file [datastore1] 9026841c-7bda-41e5-a4ac-03d0d3e37560 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.940141] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e31b1b4-9aa3-4237-8713-55018db0ab2d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.947870] env[70013]: DEBUG oslo_vmware.api [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1037.947870] env[70013]: value = "task-4231281" [ 1037.947870] env[70013]: _type = "Task" [ 1037.947870] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.963554] env[70013]: DEBUG oslo_vmware.api [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231281, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.023646] env[70013]: DEBUG nova.network.neutron [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1038.033831] env[70013]: DEBUG nova.compute.manager [req-007d673b-4503-48f0-8281-4183a38ef8d7 req-ad85c24b-aaea-4e0c-a9f8-8608ba13bde1 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received event network-vif-deleted-96be8274-aa70-44e5-8a0c-84ad7d0f9737 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1038.033925] env[70013]: INFO nova.compute.manager [req-007d673b-4503-48f0-8281-4183a38ef8d7 req-ad85c24b-aaea-4e0c-a9f8-8608ba13bde1 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Neutron deleted interface 96be8274-aa70-44e5-8a0c-84ad7d0f9737; detaching it from the instance and deleting it from the info cache [ 1038.034233] env[70013]: DEBUG nova.network.neutron [req-007d673b-4503-48f0-8281-4183a38ef8d7 req-ad85c24b-aaea-4e0c-a9f8-8608ba13bde1 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updating instance_info_cache with network_info: [{"id": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "address": "fa:16:3e:13:06:6f", "network": {"id": "65f8a088-687e-4cc7-abc0-be598d38a63e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1928781822", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.92", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eef90bcd1794b788b1340d14c6df552", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebcb7b3a-3c", "ovs_interfaceid": "ebcb7b3a-3c04-471a-b585-af59dfa574f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.143669] env[70013]: DEBUG nova.network.neutron [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1038.176405] env[70013]: DEBUG nova.network.neutron [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.192522] env[70013]: DEBUG nova.network.neutron [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.308162] env[70013]: DEBUG nova.network.neutron [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.457783] env[70013]: DEBUG oslo_vmware.api [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231281, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166164} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.458173] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.458228] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.458407] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.458583] env[70013]: INFO nova.compute.manager [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1038.458824] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1038.459033] env[70013]: DEBUG nova.compute.manager [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1038.459134] env[70013]: DEBUG nova.network.neutron [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.479883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "63c3b57c-022a-4eee-b215-6dd89142e659" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.480112] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.480306] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.480496] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.480680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.485917] env[70013]: INFO nova.compute.manager [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Terminating instance [ 1038.537331] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63542ecc-60f3-40a3-b723-8fa0db949ecc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.548348] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e2d2a6-099e-4a64-9c86-6625c11072c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.584296] env[70013]: DEBUG nova.compute.manager [req-007d673b-4503-48f0-8281-4183a38ef8d7 req-ad85c24b-aaea-4e0c-a9f8-8608ba13bde1 service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Detach interface failed, port_id=96be8274-aa70-44e5-8a0c-84ad7d0f9737, reason: Instance 181238eb-cb0c-4740-9896-ca745b53ebe8 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1038.636216] env[70013]: DEBUG nova.scheduler.client.report [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1038.655495] env[70013]: DEBUG nova.scheduler.client.report [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1038.655495] env[70013]: DEBUG nova.compute.provider_tree [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1038.669099] env[70013]: DEBUG nova.scheduler.client.report [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1038.681306] env[70013]: INFO nova.compute.manager [-] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Took 1.03 seconds to deallocate network for instance. [ 1038.690185] env[70013]: DEBUG nova.scheduler.client.report [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1038.696590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-eab0c393-4d3d-4659-8225-c82ddadb330b" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.697032] env[70013]: DEBUG nova.compute.manager [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1038.697717] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.697717] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8907495-f45b-4f9a-bea2-6bcf6d94411f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.708477] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56c7d73-d7f6-4711-b6f0-872b4f518a16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.748245] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eab0c393-4d3d-4659-8225-c82ddadb330b could not be found. [ 1038.748482] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.748684] env[70013]: INFO nova.compute.manager [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1038.748956] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1038.752667] env[70013]: DEBUG nova.compute.manager [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1038.754121] env[70013]: DEBUG nova.network.neutron [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.777385] env[70013]: DEBUG nova.network.neutron [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1038.814022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Releasing lock "refresh_cache-5661914f-ccd6-48da-bccc-a6b9065338c3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.814022] env[70013]: DEBUG nova.compute.manager [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1038.814209] env[70013]: DEBUG nova.compute.manager [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1038.814314] env[70013]: DEBUG nova.network.neutron [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.821655] env[70013]: DEBUG nova.network.neutron [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.838111] env[70013]: DEBUG nova.network.neutron [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1038.990752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.990955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquired lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1038.991185] env[70013]: DEBUG nova.network.neutron [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.124807] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cf60df-65ef-4053-bea4-68853a61f0e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.133311] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a44a1fe-81e6-44dc-b40a-463f507e909c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.165342] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54176847-28be-41c2-abad-63608efebe37 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.173939] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f021d778-1321-41e2-bad0-05ced79d5368 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.188408] env[70013]: DEBUG nova.compute.provider_tree [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.190494] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1039.232248] env[70013]: DEBUG nova.network.neutron [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.282613] env[70013]: DEBUG nova.network.neutron [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.327601] env[70013]: INFO nova.compute.manager [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Took 1.74 seconds to deallocate network for instance. [ 1039.339801] env[70013]: DEBUG nova.network.neutron [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.526596] env[70013]: DEBUG nova.network.neutron [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1039.651386] env[70013]: DEBUG nova.network.neutron [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.692767] env[70013]: DEBUG nova.scheduler.client.report [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1039.735593] env[70013]: INFO nova.compute.manager [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Took 1.28 seconds to deallocate network for instance. [ 1039.785534] env[70013]: INFO nova.compute.manager [-] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Took 1.03 seconds to deallocate network for instance. [ 1039.841022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1039.841605] env[70013]: INFO nova.compute.manager [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 5661914f-ccd6-48da-bccc-a6b9065338c3] Took 1.03 seconds to deallocate network for instance. [ 1040.063659] env[70013]: DEBUG nova.compute.manager [req-05e8cc21-9b0c-4705-aa35-d89392c0e779 req-8039426f-97b0-43b6-a9bb-e2ee6a7c81ac service nova] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Received event network-vif-deleted-ebcb7b3a-3c04-471a-b585-af59dfa574f3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1040.063881] env[70013]: DEBUG nova.compute.manager [req-05e8cc21-9b0c-4705-aa35-d89392c0e779 req-8039426f-97b0-43b6-a9bb-e2ee6a7c81ac service nova] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Received event network-vif-deleted-b3de0526-9f74-4f2b-9519-cdb229f128dd {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1040.154254] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Releasing lock "refresh_cache-63c3b57c-022a-4eee-b215-6dd89142e659" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1040.154693] env[70013]: DEBUG nova.compute.manager [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1040.154889] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.155218] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-738e3de7-129f-4092-b0e2-83f5e9d5aba5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.169650] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c1ddcf-ab70-4dc7-9fb5-dd8a77fda475 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.206072] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.206676] env[70013]: DEBUG nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1040.209777] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 63c3b57c-022a-4eee-b215-6dd89142e659 could not be found. [ 1040.210735] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1040.210735] env[70013]: INFO nova.compute.manager [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1040.210735] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1040.212100] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.569s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.213574] env[70013]: INFO nova.compute.claims [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.220314] env[70013]: DEBUG nova.compute.manager [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1040.220445] env[70013]: DEBUG nova.network.neutron [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1040.244576] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.245774] env[70013]: DEBUG nova.network.neutron [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1040.292313] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.713827] env[70013]: DEBUG nova.compute.utils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1040.715334] env[70013]: DEBUG nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1040.715532] env[70013]: DEBUG nova.network.neutron [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1040.750912] env[70013]: DEBUG nova.network.neutron [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.780026] env[70013]: DEBUG nova.policy [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ceac025cfc34c759d5bc63d2f3381aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '067cf730941645adb69270d31c369ab7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1040.884768] env[70013]: INFO nova.scheduler.client.report [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Deleted allocations for instance 5661914f-ccd6-48da-bccc-a6b9065338c3 [ 1041.152461] env[70013]: DEBUG nova.network.neutron [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Successfully created port: e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1041.219039] env[70013]: DEBUG nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1041.255358] env[70013]: INFO nova.compute.manager [-] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Took 1.03 seconds to deallocate network for instance. [ 1041.402385] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25c1c7c0-9f77-4f07-9012-5ceec3bf17c9 tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "5661914f-ccd6-48da-bccc-a6b9065338c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.668s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.727034] env[70013]: INFO nova.virt.block_device [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Booting with volume e7be3d36-ff28-48b2-936d-b40d8e49c89e at /dev/sda [ 1041.744034] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7078d093-1fab-43b0-a394-da024dad1893 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.753638] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d92e88d-ee09-4474-84e6-fa176b297291 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.789122] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1041.791371] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ecc128-10fa-4428-96be-07dd062b6161 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.794124] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37d29180-e4c2-4b2b-98f2-325e8487400a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.805123] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72baadb-52c5-4910-acf0-89c717e8191a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.811041] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6eba57a-b8de-4f91-ab06-5030bf4e4c9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.832559] env[70013]: DEBUG nova.compute.provider_tree [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.845867] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3027604-b1ea-4d9f-a085-57867a52a21f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.855904] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e07d568-c6c5-4d0a-8dad-211fe5029181 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.888857] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb281cb-fa8a-42ba-9912-767bd12c07aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.897767] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef3da84-f106-4446-8078-41e10bcfe1b0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.918159] env[70013]: DEBUG nova.virt.block_device [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updating existing volume attachment record: bf2de11b-617f-4ccf-a2fd-b5c44c12fe90 {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1042.076483] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "c57e876a-8696-4ce5-b533-7580f484414e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.076745] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "c57e876a-8696-4ce5-b533-7580f484414e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.346268] env[70013]: DEBUG nova.scheduler.client.report [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1042.579549] env[70013]: DEBUG nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1042.841830] env[70013]: DEBUG nova.compute.manager [req-ff94a345-e4e9-4ec1-9027-d3c5c74ee626 req-37c0e1be-c49c-4b5d-9279-346f1d0f6f50 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Received event network-vif-plugged-e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1042.844845] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff94a345-e4e9-4ec1-9027-d3c5c74ee626 req-37c0e1be-c49c-4b5d-9279-346f1d0f6f50 service nova] Acquiring lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.844845] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff94a345-e4e9-4ec1-9027-d3c5c74ee626 req-37c0e1be-c49c-4b5d-9279-346f1d0f6f50 service nova] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.844845] env[70013]: DEBUG oslo_concurrency.lockutils [req-ff94a345-e4e9-4ec1-9027-d3c5c74ee626 req-37c0e1be-c49c-4b5d-9279-346f1d0f6f50 service nova] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.844845] env[70013]: DEBUG nova.compute.manager [req-ff94a345-e4e9-4ec1-9027-d3c5c74ee626 req-37c0e1be-c49c-4b5d-9279-346f1d0f6f50 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] No waiting events found dispatching network-vif-plugged-e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1042.844845] env[70013]: WARNING nova.compute.manager [req-ff94a345-e4e9-4ec1-9027-d3c5c74ee626 req-37c0e1be-c49c-4b5d-9279-346f1d0f6f50 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Received unexpected event network-vif-plugged-e56ae4d8-5c16-4511-84c0-9c50ce3fa345 for instance with vm_state building and task_state block_device_mapping. [ 1042.855041] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.641s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.855041] env[70013]: DEBUG nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1042.858839] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.249s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.865109] env[70013]: INFO nova.compute.claims [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.951821] env[70013]: DEBUG nova.network.neutron [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Successfully updated port: e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.108081] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1043.374136] env[70013]: DEBUG nova.compute.utils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1043.381264] env[70013]: DEBUG nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1043.381723] env[70013]: DEBUG nova.network.neutron [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1043.459469] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Acquiring lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.459557] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Acquired lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.459727] env[70013]: DEBUG nova.network.neutron [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1043.564544] env[70013]: DEBUG nova.policy [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4c147c04d6a4b838cc27c322c389a44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9659cfe4a7a84e47af14354ffe4005cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1043.889521] env[70013]: DEBUG nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1044.024250] env[70013]: DEBUG nova.network.neutron [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1044.040078] env[70013]: DEBUG nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1044.040637] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1044.040847] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.041013] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1044.041326] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.041480] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1044.041627] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1044.041834] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1044.041988] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1044.042435] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1044.042632] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1044.042817] env[70013]: DEBUG nova.virt.hardware [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1044.043805] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a55ff1-c8c8-4579-9105-b22a6d9b9c32 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.058326] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209f8ace-49f9-448c-b3b6-71a044689389 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.224864] env[70013]: DEBUG nova.network.neutron [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Successfully created port: 38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1044.334991] env[70013]: DEBUG nova.network.neutron [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updating instance_info_cache with network_info: [{"id": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "address": "fa:16:3e:80:91:0f", "network": {"id": "dbcd8ca1-4ef5-41e3-858c-c8d8276b3764", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1649074981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "067cf730941645adb69270d31c369ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56ae4d8-5c", "ovs_interfaceid": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.396915] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf2e5c8-2f55-48db-8d20-4b49d98fd865 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.413898] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fc14bf-1b05-403d-a33e-d4e630f94a27 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.452664] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ce8f73-3ebf-423e-be6d-5a1be2875d08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.462211] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d567ad6a-17f9-4a5a-9a47-79e9d557d29b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.479693] env[70013]: DEBUG nova.compute.provider_tree [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.838772] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Releasing lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.840243] env[70013]: DEBUG nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Instance network_info: |[{"id": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "address": "fa:16:3e:80:91:0f", "network": {"id": "dbcd8ca1-4ef5-41e3-858c-c8d8276b3764", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1649074981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "067cf730941645adb69270d31c369ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56ae4d8-5c", "ovs_interfaceid": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1044.840414] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:91:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca401eaa-889a-4f9f-ac9a-56b4c41bfc06', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e56ae4d8-5c16-4511-84c0-9c50ce3fa345', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.848122] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Creating folder: Project (067cf730941645adb69270d31c369ab7). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.848435] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d3a0685-5242-4027-913a-e58837a0b833 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.865548] env[70013]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1044.865548] env[70013]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=70013) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1044.865548] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Folder already exists: Project (067cf730941645adb69270d31c369ab7). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1044.865548] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Creating folder: Instances. Parent ref: group-v837175. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.865548] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99fe0778-84b8-437f-9f2a-e80332c488ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.876768] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Created folder: Instances in parent group-v837175. [ 1044.877039] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1044.877248] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.880024] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39d38ccc-9c4f-4230-94f9-98e00b48f93e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.901062] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.901062] env[70013]: value = "task-4231285" [ 1044.901062] env[70013]: _type = "Task" [ 1044.901062] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.908224] env[70013]: DEBUG nova.compute.manager [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Received event network-changed-e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1044.908224] env[70013]: DEBUG nova.compute.manager [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Refreshing instance network info cache due to event network-changed-e56ae4d8-5c16-4511-84c0-9c50ce3fa345. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1044.908508] env[70013]: DEBUG oslo_concurrency.lockutils [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] Acquiring lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.908508] env[70013]: DEBUG oslo_concurrency.lockutils [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] Acquired lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1044.909475] env[70013]: DEBUG nova.network.neutron [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Refreshing network info cache for port e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1044.913129] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231285, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.914497] env[70013]: DEBUG nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1044.952440] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1044.954168] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.954356] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1044.954541] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.954687] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1044.954833] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1044.955066] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1044.955249] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1044.955421] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1044.955585] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1044.955762] env[70013]: DEBUG nova.virt.hardware [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1044.957446] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41bd56c-0c70-453b-b8a6-d63d082d568f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.968521] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a022e077-37e9-4ecb-b243-c48eb962fd36 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.989872] env[70013]: DEBUG nova.scheduler.client.report [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1045.415094] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231285, 'name': CreateVM_Task, 'duration_secs': 0.364152} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.415094] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.415094] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'bf2de11b-617f-4ccf-a2fd-b5c44c12fe90', 'guest_format': None, 'device_type': None, 'mount_device': '/dev/sda', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837178', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'name': 'volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'serial': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e'}, 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=70013) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1045.415094] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Root volume attach. Driver type: vmdk {{(pid=70013) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1045.417134] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2622bf-ef56-4ee0-ab65-d2a1f15f7aa8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.429028] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956ce23c-1317-42b7-a7d8-b9dabec699e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.435905] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0d1daf-86ab-413c-899e-de7a472071aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.444234] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-4d558668-153c-446e-8ce1-b7b440aa3019 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.456171] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1045.456171] env[70013]: value = "task-4231286" [ 1045.456171] env[70013]: _type = "Task" [ 1045.456171] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.467647] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231286, 'name': RelocateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.498022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.498022] env[70013]: DEBUG nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1045.500235] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.211s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.500689] env[70013]: DEBUG nova.objects.instance [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1045.748978] env[70013]: DEBUG nova.network.neutron [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updated VIF entry in instance network info cache for port e56ae4d8-5c16-4511-84c0-9c50ce3fa345. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.748978] env[70013]: DEBUG nova.network.neutron [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updating instance_info_cache with network_info: [{"id": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "address": "fa:16:3e:80:91:0f", "network": {"id": "dbcd8ca1-4ef5-41e3-858c-c8d8276b3764", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1649074981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "067cf730941645adb69270d31c369ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56ae4d8-5c", "ovs_interfaceid": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.968916] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231286, 'name': RelocateVM_Task} progress is 97%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.004017] env[70013]: DEBUG nova.network.neutron [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Successfully updated port: 38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.009961] env[70013]: DEBUG nova.compute.utils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1046.012195] env[70013]: DEBUG nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1046.012195] env[70013]: DEBUG nova.network.neutron [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1046.059844] env[70013]: DEBUG nova.policy [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3bcb0d0bd1e44eb960d676070189f27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3d42e73dc0a4a06a2022d8b54e13e12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1046.253989] env[70013]: DEBUG oslo_concurrency.lockutils [req-7ba38e73-dec4-4b54-8730-73e13778a582 req-df6d828f-14cc-4f9c-9f7b-e4d376693eb3 service nova] Releasing lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1046.341675] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.341928] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.383493] env[70013]: DEBUG nova.network.neutron [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Successfully created port: 2ac9b499-cd46-43fe-a6ca-74f684311b29 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1046.470017] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231286, 'name': RelocateVM_Task} progress is 98%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.513986] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f76a6e3-42ef-4dba-be93-b76554c88707 tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.515088] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.515161] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquired lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1046.515312] env[70013]: DEBUG nova.network.neutron [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.516325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.583s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.516545] env[70013]: DEBUG nova.objects.instance [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lazy-loading 'resources' on Instance uuid 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.520839] env[70013]: DEBUG nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1046.794511] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.794511] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.794511] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.794511] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.794511] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.798183] env[70013]: INFO nova.compute.manager [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Terminating instance [ 1046.860168] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.860476] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.860669] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.860883] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.860963] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.861554] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.861554] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1046.861554] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.946352] env[70013]: DEBUG nova.compute.manager [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Received event network-vif-plugged-38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1046.946652] env[70013]: DEBUG oslo_concurrency.lockutils [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] Acquiring lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.946767] env[70013]: DEBUG oslo_concurrency.lockutils [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.946930] env[70013]: DEBUG oslo_concurrency.lockutils [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.948332] env[70013]: DEBUG nova.compute.manager [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] No waiting events found dispatching network-vif-plugged-38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1046.948332] env[70013]: WARNING nova.compute.manager [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Received unexpected event network-vif-plugged-38ceae0a-4242-4587-9200-5cb36bbb5814 for instance with vm_state building and task_state spawning. [ 1046.948332] env[70013]: DEBUG nova.compute.manager [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Received event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1046.948332] env[70013]: DEBUG nova.compute.manager [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing instance network info cache due to event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1046.948332] env[70013]: DEBUG oslo_concurrency.lockutils [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] Acquiring lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.969876] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231286, 'name': RelocateVM_Task, 'duration_secs': 1.054766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.970223] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1046.970422] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837178', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'name': 'volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'serial': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1046.971288] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14ae21e-9581-4d53-afa8-b0a4e6599961 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.990996] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1255b5-8a0c-4fe5-9908-ae8abc6c7afc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.014444] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e/volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.015139] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-124d36fd-a558-4821-bfd6-8d9a3dfd8805 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.044958] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1047.044958] env[70013]: value = "task-4231287" [ 1047.044958] env[70013]: _type = "Task" [ 1047.044958] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.055381] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231287, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.088322] env[70013]: DEBUG nova.network.neutron [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1047.108080] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "addf8ba9-142c-4458-8586-b06e52118d71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.108385] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "addf8ba9-142c-4458-8586-b06e52118d71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.264169] env[70013]: DEBUG nova.network.neutron [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.301955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.301955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquired lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.301955] env[70013]: DEBUG nova.network.neutron [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1047.364984] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.398460] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e02df110-7add-48a6-b5fb-40642ede57e1 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.398720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e02df110-7add-48a6-b5fb-40642ede57e1 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.399158] env[70013]: DEBUG nova.objects.instance [None req-e02df110-7add-48a6-b5fb-40642ede57e1 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'flavor' on Instance uuid 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.499595] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3b0160-a7b4-430c-adb3-e9966ad75936 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.507677] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafabfa7-be8b-4bb3-9d1a-39ea414ee178 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.541712] env[70013]: DEBUG nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1047.545792] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1391cc81-78d4-4755-8425-f9904073dba3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.560493] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f865aa2-7b6e-4f3e-b629-a719114097be {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.564351] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231287, 'name': ReconfigVM_Task, 'duration_secs': 0.298022} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.566931] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Reconfigured VM instance instance-00000059 to attach disk [datastore1] volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e/volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.573034] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c234059a-c158-4164-83e0-40d921cd046d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.584773] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1047.585021] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1047.585266] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1047.585477] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1047.585627] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1047.585775] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1047.585985] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1047.586161] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1047.586331] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1047.586497] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1047.586672] env[70013]: DEBUG nova.virt.hardware [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1047.596307] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2340f180-a739-4242-9d0d-aacb4bcb08fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.599205] env[70013]: DEBUG nova.compute.provider_tree [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.609919] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817ff58c-709d-4fa4-89e8-38201af1312e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.614379] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1047.614379] env[70013]: value = "task-4231288" [ 1047.614379] env[70013]: _type = "Task" [ 1047.614379] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.618020] env[70013]: DEBUG nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1047.642367] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231288, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.767167] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Releasing lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.767584] env[70013]: DEBUG nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Instance network_info: |[{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1047.771028] env[70013]: DEBUG oslo_concurrency.lockutils [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] Acquired lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.771028] env[70013]: DEBUG nova.network.neutron [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.771028] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:c6:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '419a5b3f-4c6f-4168-9def-746b4d8c5c24', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38ceae0a-4242-4587-9200-5cb36bbb5814', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.777099] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Creating folder: Project (9659cfe4a7a84e47af14354ffe4005cb). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1047.778349] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-152f75f5-8e82-4f71-93f7-ddf64b7bded1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.789593] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Created folder: Project (9659cfe4a7a84e47af14354ffe4005cb) in parent group-v836999. [ 1047.789810] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Creating folder: Instances. Parent ref: group-v837189. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1047.789977] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b4058fe-d614-4748-b446-c2452bc81c94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.799858] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Created folder: Instances in parent group-v837189. [ 1047.800063] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1047.801358] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.801358] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef6183d9-1055-4859-84a8-71b73d35d025 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.822182] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.822182] env[70013]: value = "task-4231291" [ 1047.822182] env[70013]: _type = "Task" [ 1047.822182] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.830799] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231291, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.837058] env[70013]: DEBUG nova.network.neutron [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1047.903337] env[70013]: DEBUG nova.objects.instance [None req-e02df110-7add-48a6-b5fb-40642ede57e1 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'pci_requests' on Instance uuid 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.952290] env[70013]: DEBUG nova.network.neutron [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.013379] env[70013]: DEBUG nova.network.neutron [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Successfully updated port: 2ac9b499-cd46-43fe-a6ca-74f684311b29 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.102533] env[70013]: DEBUG nova.scheduler.client.report [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1048.131514] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231288, 'name': ReconfigVM_Task, 'duration_secs': 0.140169} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.131947] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837178', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'name': 'volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'serial': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1048.132717] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0ea3a8b-007e-43e4-a678-c66b7fb09ab1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.144912] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1048.144912] env[70013]: value = "task-4231292" [ 1048.144912] env[70013]: _type = "Task" [ 1048.144912] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.154481] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231292, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.155939] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.197734] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.198061] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.198237] env[70013]: INFO nova.compute.manager [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Shelving [ 1048.337241] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231291, 'name': CreateVM_Task, 'duration_secs': 0.391465} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.337499] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.338260] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.338536] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.340103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1048.340103] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca4d31c-4cb4-4995-a790-0e93f27527b7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.345634] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1048.345634] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]520055e7-224d-e7e2-84d2-38d33841e535" [ 1048.345634] env[70013]: _type = "Task" [ 1048.345634] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.355307] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520055e7-224d-e7e2-84d2-38d33841e535, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.406524] env[70013]: DEBUG nova.objects.base [None req-e02df110-7add-48a6-b5fb-40642ede57e1 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Object Instance<1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6> lazy-loaded attributes: flavor,pci_requests {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1048.406801] env[70013]: DEBUG nova.network.neutron [None req-e02df110-7add-48a6-b5fb-40642ede57e1 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1048.455020] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Releasing lock "refresh_cache-2c0e4326-c33a-42bb-b793-a100157b1c03" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.455493] env[70013]: DEBUG nova.compute.manager [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1048.455699] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.456035] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8298cea0-4906-44c3-94df-28c54c2c84bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.466052] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca3e4bd-03d1-45d1-a0c7-46b59739e10a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.498764] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2c0e4326-c33a-42bb-b793-a100157b1c03 could not be found. [ 1048.498764] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1048.498972] env[70013]: INFO nova.compute.manager [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1048.499206] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1048.499462] env[70013]: DEBUG nova.compute.manager [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1048.499553] env[70013]: DEBUG nova.network.neutron [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1048.518277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-ec7d0e9a-4a73-4681-9395-a3c963772f62" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.518277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-ec7d0e9a-4a73-4681-9395-a3c963772f62" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.518277] env[70013]: DEBUG nova.network.neutron [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1048.526726] env[70013]: DEBUG nova.network.neutron [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1048.582790] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e02df110-7add-48a6-b5fb-40642ede57e1 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.184s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.608500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.092s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.611404] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.041s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.611623] env[70013]: DEBUG oslo_concurrency.lockutils [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.611914] env[70013]: INFO nova.compute.manager [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] [instance: 789fffd9-6725-4bf6-9144-dd603b0a521f] Successfully reverted task state from None on failure for instance. [ 1048.614657] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.913s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.614880] env[70013]: DEBUG nova.objects.instance [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lazy-loading 'resources' on Instance uuid 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server [None req-07863f97-4bfb-4c52-bddd-937c3aa7641f tempest-ServerRescueNegativeTestJSON-1434609939 tempest-ServerRescueNegativeTestJSON-1434609939-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9cf38602-1b07-4c78-b3ce-ac9110fd60cb"}]} [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1048.619094] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1048.620766] env[70013]: ERROR oslo_messaging.rpc.server [ 1048.656159] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231292, 'name': Rename_Task, 'duration_secs': 0.160101} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.656572] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1048.656896] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3276a81-b467-4e16-b69e-87163dbd90a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.667701] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1048.667701] env[70013]: value = "task-4231293" [ 1048.667701] env[70013]: _type = "Task" [ 1048.667701] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.669123] env[70013]: INFO nova.scheduler.client.report [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleted allocations for instance 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7 [ 1048.682049] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.683665] env[70013]: DEBUG nova.network.neutron [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updated VIF entry in instance network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.684072] env[70013]: DEBUG nova.network.neutron [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.857165] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520055e7-224d-e7e2-84d2-38d33841e535, 'name': SearchDatastore_Task, 'duration_secs': 0.010321} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.857490] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.857741] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.858016] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.858176] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.858367] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.858646] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44d88483-527c-4571-baad-b717fd46768f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.868774] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.869014] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1048.870009] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-782b13b3-99ef-437d-9b2f-6983ebde8e74 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.876725] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1048.876725] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5292fb90-ec51-5f82-fb85-a65773614bfd" [ 1048.876725] env[70013]: _type = "Task" [ 1048.876725] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.886140] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5292fb90-ec51-5f82-fb85-a65773614bfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.026176] env[70013]: DEBUG nova.compute.manager [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Received event network-vif-plugged-2ac9b499-cd46-43fe-a6ca-74f684311b29 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1049.026540] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] Acquiring lock "ec7d0e9a-4a73-4681-9395-a3c963772f62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.026777] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.026999] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.027295] env[70013]: DEBUG nova.compute.manager [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] No waiting events found dispatching network-vif-plugged-2ac9b499-cd46-43fe-a6ca-74f684311b29 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1049.027470] env[70013]: WARNING nova.compute.manager [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Received unexpected event network-vif-plugged-2ac9b499-cd46-43fe-a6ca-74f684311b29 for instance with vm_state building and task_state spawning. [ 1049.027683] env[70013]: DEBUG nova.compute.manager [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Received event network-changed-2ac9b499-cd46-43fe-a6ca-74f684311b29 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1049.027883] env[70013]: DEBUG nova.compute.manager [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Refreshing instance network info cache due to event network-changed-2ac9b499-cd46-43fe-a6ca-74f684311b29. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1049.028122] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] Acquiring lock "refresh_cache-ec7d0e9a-4a73-4681-9395-a3c963772f62" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.028811] env[70013]: DEBUG nova.network.neutron [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.090549] env[70013]: DEBUG nova.network.neutron [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1049.187940] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a1640009-325c-41c5-92a3-3622a846938a tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "073f3ebf-1672-47fc-b4ed-cdb3651a8ee7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.952s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.189173] env[70013]: DEBUG oslo_concurrency.lockutils [req-9555ba40-6578-4293-98fd-0be582cc78f2 req-cbaa1c58-0393-4b1f-9277-8d93dffdb18b service nova] Releasing lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.189638] env[70013]: DEBUG oslo_vmware.api [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231293, 'name': PowerOnVM_Task, 'duration_secs': 0.48786} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.190379] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1049.190780] env[70013]: INFO nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Took 5.15 seconds to spawn the instance on the hypervisor. [ 1049.191032] env[70013]: DEBUG nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1049.191969] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ae35e0-7451-4df5-b23d-a173d3af999d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.213365] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.213694] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-440808cf-c036-4762-b515-8dad962c94cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.226209] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1049.226209] env[70013]: value = "task-4231294" [ 1049.226209] env[70013]: _type = "Task" [ 1049.226209] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.242963] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.266456] env[70013]: DEBUG nova.network.neutron [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Updating instance_info_cache with network_info: [{"id": "2ac9b499-cd46-43fe-a6ca-74f684311b29", "address": "fa:16:3e:3c:3a:67", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ac9b499-cd", "ovs_interfaceid": "2ac9b499-cd46-43fe-a6ca-74f684311b29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.397622] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5292fb90-ec51-5f82-fb85-a65773614bfd, 'name': SearchDatastore_Task, 'duration_secs': 0.010038} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.403781] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3609273-91e4-4c5b-b73b-845d85777928 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.413393] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1049.413393] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211da80-6703-77ec-7c55-f23227a6a3c3" [ 1049.413393] env[70013]: _type = "Task" [ 1049.413393] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.425272] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211da80-6703-77ec-7c55-f23227a6a3c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.531672] env[70013]: INFO nova.compute.manager [-] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Took 1.03 seconds to deallocate network for instance. [ 1049.645462] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa46fcd-d2b4-4372-810c-60724a2e9123 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.658665] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc82099-e406-4beb-92b5-cb31ec195d45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.690929] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f72e338-9ccd-4277-9bd4-35bdb4afd276 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.701341] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd17569-2470-4d69-baa7-9d18f5b51b94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.717983] env[70013]: DEBUG nova.compute.provider_tree [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.722277] env[70013]: INFO nova.compute.manager [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Took 34.94 seconds to build instance. [ 1049.742125] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231294, 'name': PowerOffVM_Task, 'duration_secs': 0.231217} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.742485] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.743841] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314b1ca3-65e8-49b9-8ab0-89b97f1e7880 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.773820] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-ec7d0e9a-4a73-4681-9395-a3c963772f62" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.774200] env[70013]: DEBUG nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Instance network_info: |[{"id": "2ac9b499-cd46-43fe-a6ca-74f684311b29", "address": "fa:16:3e:3c:3a:67", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ac9b499-cd", "ovs_interfaceid": "2ac9b499-cd46-43fe-a6ca-74f684311b29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1049.774928] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] Acquired lock "refresh_cache-ec7d0e9a-4a73-4681-9395-a3c963772f62" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1049.775112] env[70013]: DEBUG nova.network.neutron [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Refreshing network info cache for port 2ac9b499-cd46-43fe-a6ca-74f684311b29 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1049.780024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:3a:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ac9b499-cd46-43fe-a6ca-74f684311b29', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.785232] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1049.785919] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c919571f-af94-4ae0-bc94-6b7359e9a72a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.789539] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1049.789775] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c202d806-edb1-4c84-8e4e-edfe39c2b78b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.811692] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.811692] env[70013]: value = "task-4231295" [ 1049.811692] env[70013]: _type = "Task" [ 1049.811692] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.820423] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231295, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.924916] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211da80-6703-77ec-7c55-f23227a6a3c3, 'name': SearchDatastore_Task, 'duration_secs': 0.013638} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.925225] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.925490] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 8810bebc-448c-4e9b-9fbb-ce9e24611dc9/8810bebc-448c-4e9b-9fbb-ce9e24611dc9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.925746] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ffd4477d-26e2-4a0d-a8fc-f098f0c31006 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.933433] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1049.933433] env[70013]: value = "task-4231296" [ 1049.933433] env[70013]: _type = "Task" [ 1049.933433] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.942517] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231296, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.040714] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.224715] env[70013]: DEBUG nova.scheduler.client.report [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1050.230339] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4dd7ddc3-7c22-4edc-bed7-940f13c203c7 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.466s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.314011] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1050.314628] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-69c0eea8-c4a4-435d-8c91-b635b4bf0a81 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.329579] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231295, 'name': CreateVM_Task, 'duration_secs': 0.447473} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.331695] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.332474] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1050.332474] env[70013]: value = "task-4231297" [ 1050.332474] env[70013]: _type = "Task" [ 1050.332474] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.333346] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.333789] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.334014] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1050.334413] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-deb03c60-6f51-40c9-a29b-03ac3790a298 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.362609] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231297, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.363154] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1050.363154] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525be066-cd05-dd73-57da-4f7487d3e534" [ 1050.363154] env[70013]: _type = "Task" [ 1050.363154] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.374985] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525be066-cd05-dd73-57da-4f7487d3e534, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.447894] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231296, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.589831] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.589831] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.589831] env[70013]: DEBUG nova.objects.instance [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'flavor' on Instance uuid 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.701690] env[70013]: DEBUG nova.network.neutron [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Updated VIF entry in instance network info cache for port 2ac9b499-cd46-43fe-a6ca-74f684311b29. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1050.702126] env[70013]: DEBUG nova.network.neutron [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Updating instance_info_cache with network_info: [{"id": "2ac9b499-cd46-43fe-a6ca-74f684311b29", "address": "fa:16:3e:3c:3a:67", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ac9b499-cd", "ovs_interfaceid": "2ac9b499-cd46-43fe-a6ca-74f684311b29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.735158] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.120s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.738017] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.793s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.738277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.738487] env[70013]: INFO nova.compute.manager [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 9a61dedd-3764-4bd9-a300-480cc7d14a21] Successfully reverted task state from None on failure for instance. [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server [None req-c60e5809-0813-46c9-9161-2a568b0d635e tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-50ab243d-fc99-4583-8f01-ac2bfba78d9c"}]} [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1050.742687] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1050.744256] env[70013]: ERROR oslo_messaging.rpc.server [ 1050.744256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.685s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.745776] env[70013]: INFO nova.compute.claims [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.785252] env[70013]: INFO nova.scheduler.client.report [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Deleted allocations for instance 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9 [ 1050.849019] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231297, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.876850] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525be066-cd05-dd73-57da-4f7487d3e534, 'name': SearchDatastore_Task, 'duration_secs': 0.073973} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.877330] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1050.877634] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.877932] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.878138] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.878375] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.878718] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d82238d3-e091-44db-99d0-a7370ebe7f9f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.889420] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.889659] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1050.890444] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83328545-e25e-44d1-853f-1174a62cc14e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.898496] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1050.898496] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527795be-66eb-23dd-f9e4-7cca3cdf3320" [ 1050.898496] env[70013]: _type = "Task" [ 1050.898496] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.910135] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527795be-66eb-23dd-f9e4-7cca3cdf3320, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.949392] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231296, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545142} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.949894] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 8810bebc-448c-4e9b-9fbb-ce9e24611dc9/8810bebc-448c-4e9b-9fbb-ce9e24611dc9.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1050.950331] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.950851] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd4d6fcf-6c30-4daa-96f3-f15b5bb82b75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.960891] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1050.960891] env[70013]: value = "task-4231298" [ 1050.960891] env[70013]: _type = "Task" [ 1050.960891] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.971601] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231298, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.174117] env[70013]: DEBUG nova.objects.instance [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'pci_requests' on Instance uuid 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.210288] env[70013]: DEBUG oslo_concurrency.lockutils [req-f5409973-3762-432e-9c3c-6a20bd369ee2 req-a6cc5ff5-34c5-4e87-af8e-037a0b58c724 service nova] Releasing lock "refresh_cache-ec7d0e9a-4a73-4681-9395-a3c963772f62" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.251203] env[70013]: DEBUG nova.compute.manager [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Received event network-changed-e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1051.252028] env[70013]: DEBUG nova.compute.manager [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Refreshing instance network info cache due to event network-changed-e56ae4d8-5c16-4511-84c0-9c50ce3fa345. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1051.252028] env[70013]: DEBUG oslo_concurrency.lockutils [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] Acquiring lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.252028] env[70013]: DEBUG oslo_concurrency.lockutils [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] Acquired lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1051.252028] env[70013]: DEBUG nova.network.neutron [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Refreshing network info cache for port e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1051.291529] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6a35eac0-fd7c-481f-96d4-15382591c46c tempest-ServersListShow296Test-863619679 tempest-ServersListShow296Test-863619679-project-member] Lock "0c00737c-3fd4-4ec6-8b47-1ee64096e8f9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.884s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1051.345736] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231297, 'name': CreateSnapshot_Task, 'duration_secs': 0.831949} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.346019] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1051.346914] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8754f62c-57a6-409b-93de-fec0f1597b13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.412744] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527795be-66eb-23dd-f9e4-7cca3cdf3320, 'name': SearchDatastore_Task, 'duration_secs': 0.011215} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.413691] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ac48b16-736b-4245-8914-b058b889fa19 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.422492] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1051.422492] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211bf46-2c57-86aa-c84c-f60129725eca" [ 1051.422492] env[70013]: _type = "Task" [ 1051.422492] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.430845] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211bf46-2c57-86aa-c84c-f60129725eca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.474379] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231298, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071812} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.474379] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.474770] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba766475-b82a-4c7d-9220-8a442aa99e11 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.499225] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 8810bebc-448c-4e9b-9fbb-ce9e24611dc9/8810bebc-448c-4e9b-9fbb-ce9e24611dc9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.499567] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-058985b5-5121-4b89-903e-4cd58358d499 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.521452] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1051.521452] env[70013]: value = "task-4231299" [ 1051.521452] env[70013]: _type = "Task" [ 1051.521452] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.530312] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231299, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.678647] env[70013]: DEBUG nova.objects.base [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Object Instance<1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6> lazy-loaded attributes: flavor,pci_requests {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1051.678893] env[70013]: DEBUG nova.network.neutron [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1051.741328] env[70013]: DEBUG nova.policy [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1051.864865] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1051.867972] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a22070c4-a960-41c6-8b81-d33131eb3555 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.880939] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1051.880939] env[70013]: value = "task-4231300" [ 1051.880939] env[70013]: _type = "Task" [ 1051.880939] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.898742] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231300, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.933452] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5211bf46-2c57-86aa-c84c-f60129725eca, 'name': SearchDatastore_Task, 'duration_secs': 0.091038} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.936410] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.936755] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] ec7d0e9a-4a73-4681-9395-a3c963772f62/ec7d0e9a-4a73-4681-9395-a3c963772f62.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1051.937222] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a741bc6c-fb8d-4488-a39c-db239577107a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.944948] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1051.944948] env[70013]: value = "task-4231301" [ 1051.944948] env[70013]: _type = "Task" [ 1051.944948] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.964540] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231301, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.035093] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231299, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.299787] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd531042-f30f-44aa-9059-495e6fe6487e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.309253] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2805216-0694-412d-a59a-fbf3358d163c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.355572] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcc2dd1-fc25-4c37-85ec-01f95fec0919 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.365064] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c4476b-0eed-4b8a-8b0b-e1161f8c961d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.384017] env[70013]: DEBUG nova.compute.provider_tree [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1052.398292] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231300, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.456599] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231301, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.491403] env[70013]: DEBUG nova.network.neutron [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updated VIF entry in instance network info cache for port e56ae4d8-5c16-4511-84c0-9c50ce3fa345. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1052.491789] env[70013]: DEBUG nova.network.neutron [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updating instance_info_cache with network_info: [{"id": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "address": "fa:16:3e:80:91:0f", "network": {"id": "dbcd8ca1-4ef5-41e3-858c-c8d8276b3764", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1649074981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "067cf730941645adb69270d31c369ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56ae4d8-5c", "ovs_interfaceid": "e56ae4d8-5c16-4511-84c0-9c50ce3fa345", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.531039] env[70013]: DEBUG nova.network.neutron [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Successfully created port: 7b19b577-689c-4e1d-bdb0-bab33f15bd8f {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1052.537764] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231299, 'name': ReconfigVM_Task, 'duration_secs': 0.67629} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.538203] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 8810bebc-448c-4e9b-9fbb-ce9e24611dc9/8810bebc-448c-4e9b-9fbb-ce9e24611dc9.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.539078] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aaacccb9-46fc-4fe9-959f-d41cc325c8f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.548787] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1052.548787] env[70013]: value = "task-4231302" [ 1052.548787] env[70013]: _type = "Task" [ 1052.548787] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.562387] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231302, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.897040] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231300, 'name': CloneVM_Task} progress is 95%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.914434] env[70013]: ERROR nova.scheduler.client.report [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [req-63bc5da4-63b1-403d-921f-94311ded206d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-63bc5da4-63b1-403d-921f-94311ded206d"}]} [ 1052.914811] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.915420] env[70013]: ERROR nova.compute.manager [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Traceback (most recent call last): [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] yield [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] self.set_inventory_for_provider( [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-63bc5da4-63b1-403d-921f-94311ded206d"}]} [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] During handling of the above exception, another exception occurred: [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Traceback (most recent call last): [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/compute/manager.py", line 2622, in _build_and_run_instance [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] with self.rt.instance_claim(context, instance, node, allocs, [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] return f(*args, **kwargs) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 240, in instance_claim [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] self._update(elevated, cn) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] self._update_to_placement(context, compute_node, startup) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] return attempt.get(self._wrap_exception) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] raise value [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] self.reportclient.update_from_provider_tree( [ 1052.915420] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1052.916689] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] with catch_all(pd.uuid): [ 1052.916689] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1052.916689] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] self.gen.throw(typ, value, traceback) [ 1052.916689] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1052.916689] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] raise exception.ResourceProviderSyncFailed() [ 1052.916689] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1052.916689] env[70013]: ERROR nova.compute.manager [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] [ 1052.916689] env[70013]: DEBUG nova.compute.utils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1052.917343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.305s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.917535] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.917703] env[70013]: INFO nova.compute.manager [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Successfully reverted task state from None on failure for instance. [ 1052.919881] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.202s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.920106] env[70013]: DEBUG nova.objects.instance [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lazy-loading 'resources' on Instance uuid c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.921406] env[70013]: DEBUG nova.compute.manager [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Build of instance f0acce5d-d747-43b8-a5c8-ae10c10605ee was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1052.922497] env[70013]: DEBUG nova.compute.manager [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Unplugging VIFs for instance {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1052.922497] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "refresh_cache-f0acce5d-d747-43b8-a5c8-ae10c10605ee" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.922497] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "refresh_cache-f0acce5d-d747-43b8-a5c8-ae10c10605ee" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1052.922497] env[70013]: DEBUG nova.network.neutron [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server [None req-8f17da3c-c3ac-42b4-a3b9-6ba14ca53465 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-087732a9-8a3e-47f4-845e-ba5b401e31e9"}]} [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1052.924037] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1052.925462] env[70013]: ERROR oslo_messaging.rpc.server [ 1052.956454] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231301, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66357} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.956727] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] ec7d0e9a-4a73-4681-9395-a3c963772f62/ec7d0e9a-4a73-4681-9395-a3c963772f62.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.957226] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.957528] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9d9fa17-72de-49a1-a723-01b1b7dfcdbb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.964791] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1052.964791] env[70013]: value = "task-4231303" [ 1052.964791] env[70013]: _type = "Task" [ 1052.964791] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.974755] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.997257] env[70013]: DEBUG oslo_concurrency.lockutils [req-8fc0c32a-bf43-4bd1-ad65-1ddc7b3a6877 req-f2766852-4fde-4a8f-8075-4118a5b0c710 service nova] Releasing lock "refresh_cache-5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1053.060654] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231302, 'name': Rename_Task, 'duration_secs': 0.208256} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.062073] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1053.062073] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a71f57b-79b5-4d0d-a79f-24e00ebb0172 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.068315] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1053.068315] env[70013]: value = "task-4231304" [ 1053.068315] env[70013]: _type = "Task" [ 1053.068315] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.077410] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231304, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.395266] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231300, 'name': CloneVM_Task, 'duration_secs': 1.345244} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.395598] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Created linked-clone VM from snapshot [ 1053.396381] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc42a42-5911-4158-a181-2608f2af67ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.404558] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Uploading image 16d5b38a-2afd-4319-88e0-3ebfc726db58 {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1053.433365] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1053.433365] env[70013]: value = "vm-837194" [ 1053.433365] env[70013]: _type = "VirtualMachine" [ 1053.433365] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1053.435526] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-edc6b4f7-803d-4852-851a-eaf1267d132d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.444834] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease: (returnval){ [ 1053.444834] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52931714-a411-512c-a5dd-ab8f51f85d79" [ 1053.444834] env[70013]: _type = "HttpNfcLease" [ 1053.444834] env[70013]: } obtained for exporting VM: (result){ [ 1053.444834] env[70013]: value = "vm-837194" [ 1053.444834] env[70013]: _type = "VirtualMachine" [ 1053.444834] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1053.445212] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the lease: (returnval){ [ 1053.445212] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52931714-a411-512c-a5dd-ab8f51f85d79" [ 1053.445212] env[70013]: _type = "HttpNfcLease" [ 1053.445212] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1053.446220] env[70013]: DEBUG nova.scheduler.client.report [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1053.452229] env[70013]: DEBUG nova.network.neutron [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1053.457679] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1053.457679] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52931714-a411-512c-a5dd-ab8f51f85d79" [ 1053.457679] env[70013]: _type = "HttpNfcLease" [ 1053.457679] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1053.457952] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1053.457952] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52931714-a411-512c-a5dd-ab8f51f85d79" [ 1053.457952] env[70013]: _type = "HttpNfcLease" [ 1053.457952] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1053.458727] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af76ec80-59fb-4569-92cb-18b57ac253d4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.462705] env[70013]: DEBUG nova.scheduler.client.report [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1053.462943] env[70013]: DEBUG nova.compute.provider_tree [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1053.472381] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5272362c-5ee3-db3a-8e40-de3b5a280d18/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1053.472591] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5272362c-5ee3-db3a-8e40-de3b5a280d18/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1053.475084] env[70013]: DEBUG nova.scheduler.client.report [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1053.480882] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117469} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.532435] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.533709] env[70013]: DEBUG nova.scheduler.client.report [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1053.538509] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5edf0b6-f4c3-4b91-a178-a894ee3b949f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.565987] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] ec7d0e9a-4a73-4681-9395-a3c963772f62/ec7d0e9a-4a73-4681-9395-a3c963772f62.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.569039] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b35bb58d-5107-4930-9b8a-9537ea14b72f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.584470] env[70013]: DEBUG nova.network.neutron [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.595751] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231304, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.597485] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1053.597485] env[70013]: value = "task-4231306" [ 1053.597485] env[70013]: _type = "Task" [ 1053.597485] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.615819] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8d07f2ff-37d8-4bbe-a508-7a443f1f5c9b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.033292] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e408fd-1db1-45f2-abb7-8e820fd7086f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.045646] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76b8fca-0e8a-425c-b42b-f2533d4f8f5c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.082486] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f25ba0a-3ffc-4fce-9eec-878f4688fccf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.090329] env[70013]: DEBUG nova.network.neutron [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Successfully updated port: 7b19b577-689c-4e1d-bdb0-bab33f15bd8f {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1054.091310] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "refresh_cache-f0acce5d-d747-43b8-a5c8-ae10c10605ee" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1054.092387] env[70013]: DEBUG nova.compute.manager [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=70013) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1054.092387] env[70013]: DEBUG nova.compute.manager [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1054.092387] env[70013]: DEBUG nova.network.neutron [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1054.100391] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0ef32f-c4f9-4955-809c-e7ea69e4718a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.113908] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231304, 'name': PowerOnVM_Task} progress is 71%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.115223] env[70013]: DEBUG nova.network.neutron [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1054.137031] env[70013]: DEBUG nova.compute.provider_tree [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1054.138928] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.321410] env[70013]: DEBUG nova.compute.manager [req-b8b431a3-0a7a-435c-b0e4-56b143b866f3 req-d61d44c5-fb86-43b8-a78b-63c4ac289184 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-vif-plugged-7b19b577-689c-4e1d-bdb0-bab33f15bd8f {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1054.321713] env[70013]: DEBUG oslo_concurrency.lockutils [req-b8b431a3-0a7a-435c-b0e4-56b143b866f3 req-d61d44c5-fb86-43b8-a78b-63c4ac289184 service nova] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.322340] env[70013]: DEBUG oslo_concurrency.lockutils [req-b8b431a3-0a7a-435c-b0e4-56b143b866f3 req-d61d44c5-fb86-43b8-a78b-63c4ac289184 service nova] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.322782] env[70013]: DEBUG oslo_concurrency.lockutils [req-b8b431a3-0a7a-435c-b0e4-56b143b866f3 req-d61d44c5-fb86-43b8-a78b-63c4ac289184 service nova] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.323190] env[70013]: DEBUG nova.compute.manager [req-b8b431a3-0a7a-435c-b0e4-56b143b866f3 req-d61d44c5-fb86-43b8-a78b-63c4ac289184 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] No waiting events found dispatching network-vif-plugged-7b19b577-689c-4e1d-bdb0-bab33f15bd8f {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1054.323604] env[70013]: WARNING nova.compute.manager [req-b8b431a3-0a7a-435c-b0e4-56b143b866f3 req-d61d44c5-fb86-43b8-a78b-63c4ac289184 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received unexpected event network-vif-plugged-7b19b577-689c-4e1d-bdb0-bab33f15bd8f for instance with vm_state active and task_state None. [ 1054.596014] env[70013]: DEBUG oslo_vmware.api [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231304, 'name': PowerOnVM_Task, 'duration_secs': 1.242973} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.596606] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1054.597206] env[70013]: INFO nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Took 9.68 seconds to spawn the instance on the hypervisor. [ 1054.597206] env[70013]: DEBUG nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1054.597752] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71fa44c-0fe1-4f86-a105-188b4e4f240e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.600925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.601306] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1054.601987] env[70013]: DEBUG nova.network.neutron [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.617472] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231306, 'name': ReconfigVM_Task, 'duration_secs': 0.757821} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.617987] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Reconfigured VM instance instance-0000005b to attach disk [datastore1] ec7d0e9a-4a73-4681-9395-a3c963772f62/ec7d0e9a-4a73-4681-9395-a3c963772f62.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.618969] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27f0ae2d-1dba-425e-9344-e3696ac4de67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.621167] env[70013]: DEBUG nova.network.neutron [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.631328] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1054.631328] env[70013]: value = "task-4231307" [ 1054.631328] env[70013]: _type = "Task" [ 1054.631328] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.646462] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231307, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.666133] env[70013]: ERROR nova.scheduler.client.report [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [req-92fcba01-94d0-4155-8c2b-93d8453ebad6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-92fcba01-94d0-4155-8c2b-93d8453ebad6"}]} [ 1054.666557] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.747s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.667251] env[70013]: ERROR nova.compute.manager [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Traceback (most recent call last): [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] yield [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self.set_inventory_for_provider( [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-92fcba01-94d0-4155-8c2b-93d8453ebad6"}]} [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] During handling of the above exception, another exception occurred: [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Traceback (most recent call last): [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self._delete_instance(context, instance, bdms) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self._complete_deletion(context, instance) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self._update_resource_tracker(context, instance) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self.rt.update_usage(context, instance, instance.node) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] return f(*args, **kwargs) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self._update_to_placement(context, compute_node, startup) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] return attempt.get(self._wrap_exception) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1054.667251] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] raise value [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self.reportclient.update_from_provider_tree( [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] with catch_all(pd.uuid): [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] self.gen.throw(typ, value, traceback) [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] raise exception.ResourceProviderSyncFailed() [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1054.668480] env[70013]: ERROR nova.compute.manager [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] [ 1054.670439] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.343s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.670791] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.671050] env[70013]: INFO nova.compute.manager [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] [instance: 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff] Successfully reverted task state from None on failure for instance. [ 1054.673545] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.483s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.673802] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.676661] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.836s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.677021] env[70013]: DEBUG nova.objects.instance [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lazy-loading 'resources' on Instance uuid 181238eb-cb0c-4740-9896-ca745b53ebe8 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server [None req-d55536c2-1650-47e0-ae8d-68c9cb7d1af3 tempest-ServersV294TestFqdnHostnames-1142748138 tempest-ServersV294TestFqdnHostnames-1142748138-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f23150ca-7490-47b1-8a51-526f1e183fe9"}]} [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1054.679238] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1054.680727] env[70013]: ERROR oslo_messaging.rpc.server [ 1054.704351] env[70013]: INFO nova.scheduler.client.report [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted allocations for instance 8dbef727-13e9-4231-95a8-65d015ee13be [ 1055.127052] env[70013]: INFO nova.compute.manager [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: f0acce5d-d747-43b8-a5c8-ae10c10605ee] Took 1.03 seconds to deallocate network for instance. [ 1055.130996] env[70013]: INFO nova.compute.manager [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Took 38.50 seconds to build instance. [ 1055.143506] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231307, 'name': Rename_Task, 'duration_secs': 0.225457} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.143914] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1055.146965] env[70013]: WARNING nova.network.neutron [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] b160f70a-0663-495b-bd35-440c047157c8 already exists in list: networks containing: ['b160f70a-0663-495b-bd35-440c047157c8']. ignoring it [ 1055.149142] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-081b4849-66b8-4b45-9181-c6e16b29f961 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.160811] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1055.160811] env[70013]: value = "task-4231308" [ 1055.160811] env[70013]: _type = "Task" [ 1055.160811] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.174156] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231308, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.181192] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.901s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.203510] env[70013]: DEBUG nova.scheduler.client.report [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1055.214297] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fbcc090f-4497-4eac-ae81-779db2e69a3c tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "8dbef727-13e9-4231-95a8-65d015ee13be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.519s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.225464] env[70013]: DEBUG nova.scheduler.client.report [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1055.225675] env[70013]: DEBUG nova.compute.provider_tree [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1055.242360] env[70013]: DEBUG nova.scheduler.client.report [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1055.270167] env[70013]: DEBUG nova.scheduler.client.report [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1055.487564] env[70013]: DEBUG nova.network.neutron [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "address": "fa:16:3e:2a:49:e7", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b19b577-68", "ovs_interfaceid": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.638036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7e72d1c7-4af9-4fe8-978a-78a31bba7450 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.022s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.675314] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231308, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.711499] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be82dace-d622-4228-8344-8d96b66f0072 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.722285] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4700c71b-e34a-4779-ae74-4b3acad0777a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.761496] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5970d3-4e1e-40d7-9486-f1e82d8773d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.771460] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602c4d4e-eb6b-4746-818a-a54fc2a611bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.792637] env[70013]: DEBUG nova.compute.provider_tree [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1055.990293] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1055.991186] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.991316] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1055.992216] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba416c16-5ff1-4f4d-81aa-a02dfdc99e5a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.011814] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1056.012057] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.012231] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1056.012409] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.012552] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1056.012693] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1056.012896] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1056.013071] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1056.013241] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1056.013405] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1056.013578] env[70013]: DEBUG nova.virt.hardware [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1056.023812] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfiguring VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1056.024310] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7d44d60-a3cc-4aaf-a76b-9f27af98b8a3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.047133] env[70013]: DEBUG oslo_vmware.api [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1056.047133] env[70013]: value = "task-4231309" [ 1056.047133] env[70013]: _type = "Task" [ 1056.047133] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.059692] env[70013]: DEBUG oslo_vmware.api [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231309, 'name': ReconfigVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.167202] env[70013]: INFO nova.scheduler.client.report [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted allocations for instance f0acce5d-d747-43b8-a5c8-ae10c10605ee [ 1056.189399] env[70013]: DEBUG oslo_vmware.api [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231308, 'name': PowerOnVM_Task, 'duration_secs': 0.664078} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.189399] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1056.189399] env[70013]: INFO nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Took 8.65 seconds to spawn the instance on the hypervisor. [ 1056.189399] env[70013]: DEBUG nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1056.189399] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d62ddd-8b36-484b-b75f-7c4bdee8f066 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.323695] env[70013]: ERROR nova.scheduler.client.report [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [req-38840846-5423-4067-8d5a-02f68e981ac0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-38840846-5423-4067-8d5a-02f68e981ac0"}]} [ 1056.324198] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.325246] env[70013]: ERROR nova.compute.manager [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Traceback (most recent call last): [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] yield [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self.set_inventory_for_provider( [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-38840846-5423-4067-8d5a-02f68e981ac0"}]} [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] During handling of the above exception, another exception occurred: [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Traceback (most recent call last): [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self._delete_instance(context, instance, bdms) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self._complete_deletion(context, instance) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self._update_resource_tracker(context, instance) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self.rt.update_usage(context, instance, instance.node) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] return f(*args, **kwargs) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self._update_to_placement(context, compute_node, startup) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] return attempt.get(self._wrap_exception) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1056.325246] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] raise value [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self.reportclient.update_from_provider_tree( [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] with catch_all(pd.uuid): [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] self.gen.throw(typ, value, traceback) [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] raise exception.ResourceProviderSyncFailed() [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1056.326489] env[70013]: ERROR nova.compute.manager [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] [ 1056.327426] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.083s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.329853] env[70013]: DEBUG nova.objects.instance [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'resources' on Instance uuid 9026841c-7bda-41e5-a4ac-03d0d3e37560 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.376993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "54d39dfd-6429-4009-8d26-22c1fa46243e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.376993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.376993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "54d39dfd-6429-4009-8d26-22c1fa46243e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.376993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.376993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.378355] env[70013]: INFO nova.compute.manager [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Terminating instance [ 1056.513374] env[70013]: DEBUG nova.compute.manager [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Received event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1056.513374] env[70013]: DEBUG nova.compute.manager [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing instance network info cache due to event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1056.513675] env[70013]: DEBUG oslo_concurrency.lockutils [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] Acquiring lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.513879] env[70013]: DEBUG oslo_concurrency.lockutils [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] Acquired lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.514040] env[70013]: DEBUG nova.network.neutron [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.560726] env[70013]: DEBUG oslo_vmware.api [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231309, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.688710] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c9eae966-26b4-4c74-a95c-74ce26b5d78c tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "f0acce5d-d747-43b8-a5c8-ae10c10605ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.659s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.702714] env[70013]: DEBUG nova.compute.manager [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-changed-7b19b577-689c-4e1d-bdb0-bab33f15bd8f {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1056.702979] env[70013]: DEBUG nova.compute.manager [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing instance network info cache due to event network-changed-7b19b577-689c-4e1d-bdb0-bab33f15bd8f. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1056.703323] env[70013]: DEBUG oslo_concurrency.lockutils [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.703518] env[70013]: DEBUG oslo_concurrency.lockutils [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.703721] env[70013]: DEBUG nova.network.neutron [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing network info cache for port 7b19b577-689c-4e1d-bdb0-bab33f15bd8f {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.706453] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.717147] env[70013]: INFO nova.compute.manager [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Took 29.13 seconds to build instance. [ 1056.831952] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.916s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.855937] env[70013]: DEBUG nova.scheduler.client.report [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1056.875953] env[70013]: DEBUG nova.scheduler.client.report [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1056.876209] env[70013]: DEBUG nova.compute.provider_tree [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1056.883150] env[70013]: DEBUG nova.compute.manager [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1056.883442] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.884454] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985d0277-140c-4d01-b07e-ef17ba81a2f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.892194] env[70013]: DEBUG nova.scheduler.client.report [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1056.897804] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.898207] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df466361-460c-4087-8b9f-3e9c7b03eeba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.905548] env[70013]: DEBUG oslo_vmware.api [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 1056.905548] env[70013]: value = "task-4231310" [ 1056.905548] env[70013]: _type = "Task" [ 1056.905548] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.915561] env[70013]: DEBUG oslo_vmware.api [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.916801] env[70013]: DEBUG nova.scheduler.client.report [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1057.067076] env[70013]: DEBUG oslo_vmware.api [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231309, 'name': ReconfigVM_Task, 'duration_secs': 0.903662} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.070301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.070663] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfigured VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1057.220285] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a2d7483e-7efe-404d-9d52-8bdc08648e3e tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.664s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.322294] env[70013]: DEBUG oslo_concurrency.lockutils [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.322638] env[70013]: DEBUG oslo_concurrency.lockutils [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.322828] env[70013]: DEBUG nova.compute.manager [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1057.324139] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a8d899-c408-4b27-b7aa-30d51d3a9b2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.342375] env[70013]: DEBUG nova.compute.manager [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1057.342980] env[70013]: DEBUG nova.objects.instance [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'flavor' on Instance uuid ec7d0e9a-4a73-4681-9395-a3c963772f62 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.412639] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8077b843-820a-4073-b372-08d5a8dd71ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.418702] env[70013]: DEBUG oslo_vmware.api [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231310, 'name': PowerOffVM_Task, 'duration_secs': 0.202067} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.419429] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.419599] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.419867] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3cc2505-a125-42e8-85ca-06c47179ceb6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.424999] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dfdabf-57a1-43e3-9b88-ff24f57255e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.462180] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b3696b-1640-45b9-ac7d-169b100efa10 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.465934] env[70013]: DEBUG nova.network.neutron [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updated VIF entry in instance network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.466360] env[70013]: DEBUG nova.network.neutron [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.475331] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d37e9ac-3e51-48c8-8cfb-d79e14151be1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.496557] env[70013]: DEBUG nova.compute.provider_tree [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1057.504782] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.504866] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.505093] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Deleting the datastore file [datastore2] 54d39dfd-6429-4009-8d26-22c1fa46243e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.505408] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-141b0578-d5c9-4e12-b467-22d3053c240a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.516237] env[70013]: DEBUG oslo_vmware.api [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for the task: (returnval){ [ 1057.516237] env[70013]: value = "task-4231312" [ 1057.516237] env[70013]: _type = "Task" [ 1057.516237] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.526990] env[70013]: DEBUG oslo_vmware.api [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.575943] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5120a384-43b4-42fe-ab3b-43a370934a7e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.988s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.590979] env[70013]: DEBUG nova.network.neutron [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updated VIF entry in instance network info cache for port 7b19b577-689c-4e1d-bdb0-bab33f15bd8f. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.591456] env[70013]: DEBUG nova.network.neutron [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "address": "fa:16:3e:2a:49:e7", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b19b577-68", "ovs_interfaceid": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.969713] env[70013]: DEBUG oslo_concurrency.lockutils [req-73d82c39-8ce5-4be4-9e99-d6f42936036c req-b8125906-8204-4b72-a191-cd7969e7ce73 service nova] Releasing lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.022938] env[70013]: ERROR nova.scheduler.client.report [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [req-c070cdc0-931e-46e4-9c1f-fe292e9e6ad8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc0a4c9d-b611-453e-a900-41e280c29b95. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c070cdc0-931e-46e4-9c1f-fe292e9e6ad8"}]} [ 1058.023332] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.024429] env[70013]: ERROR nova.compute.manager [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Traceback (most recent call last): [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] yield [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self.set_inventory_for_provider( [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c070cdc0-931e-46e4-9c1f-fe292e9e6ad8"}]} [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] During handling of the above exception, another exception occurred: [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Traceback (most recent call last): [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self._delete_instance(context, instance, bdms) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self._complete_deletion(context, instance) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self._update_resource_tracker(context, instance) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self.rt.update_usage(context, instance, instance.node) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] return f(*args, **kwargs) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self._update(context.elevated(), self.compute_nodes[nodename]) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self._update_to_placement(context, compute_node, startup) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] return attempt.get(self._wrap_exception) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] six.reraise(self.value[0], self.value[1], self.value[2]) [ 1058.024429] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] raise value [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self.reportclient.update_from_provider_tree( [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] with catch_all(pd.uuid): [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] self.gen.throw(typ, value, traceback) [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] raise exception.ResourceProviderSyncFailed() [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1058.025478] env[70013]: ERROR nova.compute.manager [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] [ 1058.026852] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.735s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.027512] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.031544] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.240s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.031544] env[70013]: DEBUG nova.objects.instance [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lazy-loading 'resources' on Instance uuid 63c3b57c-022a-4eee-b215-6dd89142e659 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.034055] env[70013]: DEBUG oslo_vmware.api [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Task: {'id': task-4231312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32811} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.034598] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.034598] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1058.035280] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1058.035280] env[70013]: INFO nova.compute.manager [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1058.035280] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1058.035280] env[70013]: DEBUG nova.compute.manager [-] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1058.035435] env[70013]: DEBUG nova.network.neutron [-] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1058.070947] env[70013]: INFO nova.scheduler.client.report [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted allocations for instance eab0c393-4d3d-4659-8225-c82ddadb330b [ 1058.095328] env[70013]: DEBUG oslo_concurrency.lockutils [req-118c6cd8-1996-4de4-9af4-665a71cf8404 req-45330b70-1fbf-4af8-9623-05d5d260fc06 service nova] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.356062] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1058.357074] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.357403] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9966a792-d4be-429c-948d-d06209d468c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.371836] env[70013]: DEBUG oslo_vmware.api [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1058.371836] env[70013]: value = "task-4231313" [ 1058.371836] env[70013]: _type = "Task" [ 1058.371836] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.385761] env[70013]: DEBUG oslo_vmware.api [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.535978] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.709s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.562972] env[70013]: DEBUG nova.scheduler.client.report [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1058.582885] env[70013]: DEBUG oslo_concurrency.lockutils [None req-eeec97b6-8012-4353-9555-817c8eb87dcb tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "eab0c393-4d3d-4659-8225-c82ddadb330b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.607s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.591995] env[70013]: DEBUG nova.scheduler.client.report [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1058.592268] env[70013]: DEBUG nova.compute.provider_tree [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1058.609786] env[70013]: DEBUG nova.scheduler.client.report [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1058.634756] env[70013]: DEBUG nova.scheduler.client.report [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1058.889730] env[70013]: DEBUG oslo_vmware.api [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231313, 'name': PowerOffVM_Task, 'duration_secs': 0.338637} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.890113] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1058.890401] env[70013]: DEBUG nova.compute.manager [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1058.891284] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6def70dd-f90f-405c-8a82-c982382136f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.922382] env[70013]: DEBUG nova.network.neutron [-] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.955530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "2070989d-6c90-4eb1-8508-7587ed9659d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.955866] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.956121] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "2070989d-6c90-4eb1-8508-7587ed9659d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.956367] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.956546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.958662] env[70013]: INFO nova.compute.manager [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Terminating instance [ 1058.968809] env[70013]: DEBUG nova.compute.manager [req-7f0c65ac-4de9-4cc2-997a-db12e1e9e70e req-ce583ad6-e389-401f-ac11-df9c9e9b2d39 service nova] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Received event network-vif-deleted-20f2f376-c1c7-4171-bc55-1ecdb0cc5d14 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1059.103248] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f58f777-3ff8-413c-9d01-6f9735726799 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.112573] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38a57b3-6c2e-4298-bde9-0ee56cee4d3c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.145487] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e238c689-e8b6-4366-a3f7-003c5eb7c607 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.154338] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc31056-b19c-427a-ab8b-d5a70dd01357 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.169814] env[70013]: DEBUG nova.compute.provider_tree [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.176979] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.177247] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.404154] env[70013]: DEBUG oslo_concurrency.lockutils [None req-88d9c0ba-cce8-4541-a483-e80608a89390 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.081s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1059.425460] env[70013]: INFO nova.compute.manager [-] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Took 1.39 seconds to deallocate network for instance. [ 1059.467538] env[70013]: DEBUG nova.compute.manager [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1059.467538] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.467538] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3baa79-f49f-4c12-a24e-4fd5978b2b67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.476875] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.477716] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51205c38-e32e-487e-a57f-6fe6b42cae88 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.486509] env[70013]: DEBUG oslo_vmware.api [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1059.486509] env[70013]: value = "task-4231314" [ 1059.486509] env[70013]: _type = "Task" [ 1059.486509] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.494663] env[70013]: DEBUG oslo_vmware.api [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.565500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "181238eb-cb0c-4740-9896-ca745b53ebe8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.565500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.565500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.565500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.566225] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1059.569963] env[70013]: INFO nova.compute.manager [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Terminating instance [ 1059.673940] env[70013]: DEBUG nova.scheduler.client.report [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1059.682343] env[70013]: DEBUG nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1059.757605] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.757902] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.758318] env[70013]: DEBUG nova.objects.instance [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'flavor' on Instance uuid 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.933185] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.997749] env[70013]: DEBUG oslo_vmware.api [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231314, 'name': PowerOffVM_Task, 'duration_secs': 0.227389} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.998125] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.998349] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.998675] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdd924d1-3dfd-415a-909d-98cb58bc909e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.056255] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.076302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.076536] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquired lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.076765] env[70013]: DEBUG nova.network.neutron [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1060.078269] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1060.078553] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1060.078817] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleting the datastore file [datastore2] 2070989d-6c90-4eb1-8508-7587ed9659d6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.079726] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cb49d33-aef2-4ab4-9815-7fcb5ee2a229 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.087086] env[70013]: DEBUG oslo_vmware.api [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1060.087086] env[70013]: value = "task-4231316" [ 1060.087086] env[70013]: _type = "Task" [ 1060.087086] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.098916] env[70013]: DEBUG oslo_vmware.api [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.180459] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.151s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.183935] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.076s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.186118] env[70013]: INFO nova.compute.claims [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.207263] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.208265] env[70013]: INFO nova.scheduler.client.report [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Deleted allocations for instance 63c3b57c-022a-4eee-b215-6dd89142e659 [ 1060.368516] env[70013]: DEBUG nova.objects.instance [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'pci_requests' on Instance uuid 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.598811] env[70013]: DEBUG oslo_vmware.api [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325898} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.599217] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.600033] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.600033] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.600033] env[70013]: INFO nova.compute.manager [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1060.600267] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1060.600369] env[70013]: DEBUG nova.compute.manager [-] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1060.600476] env[70013]: DEBUG nova.network.neutron [-] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1060.615244] env[70013]: DEBUG nova.network.neutron [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1060.717175] env[70013]: DEBUG oslo_concurrency.lockutils [None req-748b65c9-aed3-4fef-a0fa-a59b195f4c83 tempest-ImagesTestJSON-196229908 tempest-ImagesTestJSON-196229908-project-member] Lock "63c3b57c-022a-4eee-b215-6dd89142e659" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.237s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.744012] env[70013]: DEBUG nova.network.neutron [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.861130] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.861519] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.861739] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "ec7d0e9a-4a73-4681-9395-a3c963772f62-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.861974] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.862121] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.864355] env[70013]: INFO nova.compute.manager [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Terminating instance [ 1060.871056] env[70013]: DEBUG nova.objects.base [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Object Instance<1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6> lazy-loaded attributes: flavor,pci_requests {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1060.871056] env[70013]: DEBUG nova.network.neutron [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1060.969356] env[70013]: DEBUG nova.policy [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1061.071116] env[70013]: DEBUG nova.compute.manager [req-22b4854e-34c1-4c69-b0fb-12667ac02354 req-cab08534-6267-43cd-b526-4ff449b22c8c service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Received event network-vif-deleted-a9940c14-e6da-4fa5-b093-ba598b0bdc93 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1061.071976] env[70013]: INFO nova.compute.manager [req-22b4854e-34c1-4c69-b0fb-12667ac02354 req-cab08534-6267-43cd-b526-4ff449b22c8c service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Neutron deleted interface a9940c14-e6da-4fa5-b093-ba598b0bdc93; detaching it from the instance and deleting it from the info cache [ 1061.072491] env[70013]: DEBUG nova.network.neutron [req-22b4854e-34c1-4c69-b0fb-12667ac02354 req-cab08534-6267-43cd-b526-4ff449b22c8c service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.248497] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Releasing lock "refresh_cache-181238eb-cb0c-4740-9896-ca745b53ebe8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1061.248497] env[70013]: DEBUG nova.compute.manager [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1061.248822] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.252021] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e29e6b99-3aab-4980-997d-d01c079eb770 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.261191] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d249c240-c9b0-4327-869c-47dfecccda42 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.298311] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 181238eb-cb0c-4740-9896-ca745b53ebe8 could not be found. [ 1061.298855] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.298855] env[70013]: INFO nova.compute.manager [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1061.299278] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1061.302378] env[70013]: DEBUG nova.compute.manager [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1061.302541] env[70013]: DEBUG nova.network.neutron [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1061.322048] env[70013]: DEBUG nova.network.neutron [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1061.369329] env[70013]: DEBUG nova.compute.manager [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1061.369922] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.372161] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2eebca-2fa0-4660-98f3-112c52672319 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.384025] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.384025] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7167db25-1490-4a6b-ba27-36bfb2ba8b65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.466274] env[70013]: DEBUG nova.network.neutron [-] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.467876] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.468058] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.470294] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleting the datastore file [datastore1] ec7d0e9a-4a73-4681-9395-a3c963772f62 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.470845] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4699bb5f-ff00-457e-95e3-7d2a2914f0aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.485668] env[70013]: DEBUG oslo_vmware.api [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1061.485668] env[70013]: value = "task-4231318" [ 1061.485668] env[70013]: _type = "Task" [ 1061.485668] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.496467] env[70013]: DEBUG oslo_vmware.api [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.575092] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-13067794-e715-4e0b-93d5-11db8e93eb6d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.585969] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2906178-d1b1-4e63-8f98-131e08810e88 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.626252] env[70013]: DEBUG nova.compute.manager [req-22b4854e-34c1-4c69-b0fb-12667ac02354 req-cab08534-6267-43cd-b526-4ff449b22c8c service nova] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Detach interface failed, port_id=a9940c14-e6da-4fa5-b093-ba598b0bdc93, reason: Instance 2070989d-6c90-4eb1-8508-7587ed9659d6 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1061.655124] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc163bc4-7356-4524-bfa5-d645dd46b13e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.664103] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de2fb4f-f062-4252-9a37-be9e365b5456 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.698213] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fddf3dd-533e-4806-bc4b-c9e4c7966c39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.707736] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3107b5-12d9-41e7-8419-57314a7214db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.722975] env[70013]: DEBUG nova.compute.provider_tree [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.801690] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5272362c-5ee3-db3a-8e40-de3b5a280d18/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1061.802386] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e389443c-8dcb-4332-8b07-8f544aca490b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.809894] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5272362c-5ee3-db3a-8e40-de3b5a280d18/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1061.810391] env[70013]: ERROR oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5272362c-5ee3-db3a-8e40-de3b5a280d18/disk-0.vmdk due to incomplete transfer. [ 1061.810391] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d60cf713-edbb-4632-afb6-ed468e2dc7e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.818868] env[70013]: DEBUG oslo_vmware.rw_handles [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5272362c-5ee3-db3a-8e40-de3b5a280d18/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1061.819100] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Uploaded image 16d5b38a-2afd-4319-88e0-3ebfc726db58 to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1061.821607] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1061.821928] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-29567a3a-1b3b-49bb-b773-b6cced4bb5f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.823772] env[70013]: DEBUG nova.network.neutron [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.830458] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1061.830458] env[70013]: value = "task-4231319" [ 1061.830458] env[70013]: _type = "Task" [ 1061.830458] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.840011] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231319, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.971513] env[70013]: INFO nova.compute.manager [-] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Took 1.37 seconds to deallocate network for instance. [ 1061.996431] env[70013]: DEBUG oslo_vmware.api [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258378} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.996744] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.996939] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.997133] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.997314] env[70013]: INFO nova.compute.manager [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1061.997690] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1061.997902] env[70013]: DEBUG nova.compute.manager [-] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1061.997999] env[70013]: DEBUG nova.network.neutron [-] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1062.226699] env[70013]: DEBUG nova.scheduler.client.report [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1062.331748] env[70013]: INFO nova.compute.manager [-] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Took 1.03 seconds to deallocate network for instance. [ 1062.351813] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231319, 'name': Destroy_Task, 'duration_secs': 0.369966} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.351813] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Destroyed the VM [ 1062.351813] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1062.352929] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5f735e96-14c5-4d5e-b4e5-ad93ec9288e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.359174] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1062.359174] env[70013]: value = "task-4231320" [ 1062.359174] env[70013]: _type = "Task" [ 1062.359174] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.369302] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231320, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.478619] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.524127] env[70013]: DEBUG nova.network.neutron [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Successfully updated port: 2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1062.733398] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.733553] env[70013]: DEBUG nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1062.736377] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.371s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.736514] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.736619] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1062.737431] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.582s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.739017] env[70013]: INFO nova.compute.claims [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.742498] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a79c856-e2af-4e09-8f9b-63bc6549ba18 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.751687] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e1d244-269a-481f-aa72-5c713d6c754b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.769990] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2193460-5ef0-45d3-8b87-4d5db75096e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.775931] env[70013]: DEBUG nova.network.neutron [-] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.779408] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdeb35b-7f7d-4974-bc0d-b667b58e0106 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.812853] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178803MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1062.813046] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.843615] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.872368] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231320, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.034230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.034230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.034230] env[70013]: DEBUG nova.network.neutron [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1063.206793] env[70013]: DEBUG nova.compute.manager [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Received event network-vif-deleted-2ac9b499-cd46-43fe-a6ca-74f684311b29 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1063.207235] env[70013]: DEBUG nova.compute.manager [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-vif-plugged-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1063.207745] env[70013]: DEBUG oslo_concurrency.lockutils [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1063.207745] env[70013]: DEBUG oslo_concurrency.lockutils [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.207915] env[70013]: DEBUG oslo_concurrency.lockutils [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.207993] env[70013]: DEBUG nova.compute.manager [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] No waiting events found dispatching network-vif-plugged-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1063.208175] env[70013]: WARNING nova.compute.manager [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received unexpected event network-vif-plugged-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 for instance with vm_state active and task_state None. [ 1063.208346] env[70013]: DEBUG nova.compute.manager [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-changed-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1063.208498] env[70013]: DEBUG nova.compute.manager [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing instance network info cache due to event network-changed-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1063.208668] env[70013]: DEBUG oslo_concurrency.lockutils [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.243773] env[70013]: DEBUG nova.compute.utils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1063.245364] env[70013]: DEBUG nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1063.245580] env[70013]: DEBUG nova.network.neutron [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1063.279576] env[70013]: INFO nova.compute.manager [-] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Took 1.28 seconds to deallocate network for instance. [ 1063.298604] env[70013]: DEBUG nova.policy [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed0f82552bf43268d93688820492e44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c6094eeb5ca48dfbfbb10bc71feac67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1063.370172] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231320, 'name': RemoveSnapshot_Task, 'duration_secs': 0.578077} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.370559] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1063.371176] env[70013]: DEBUG nova.compute.manager [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1063.371780] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682ce1a1-c4cd-43e0-bad4-8022c40ce3ae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.588151] env[70013]: WARNING nova.network.neutron [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] b160f70a-0663-495b-bd35-440c047157c8 already exists in list: networks containing: ['b160f70a-0663-495b-bd35-440c047157c8']. ignoring it [ 1063.588385] env[70013]: WARNING nova.network.neutron [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] b160f70a-0663-495b-bd35-440c047157c8 already exists in list: networks containing: ['b160f70a-0663-495b-bd35-440c047157c8']. ignoring it [ 1063.640290] env[70013]: DEBUG nova.network.neutron [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Successfully created port: f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.749708] env[70013]: DEBUG nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1063.786833] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1063.885810] env[70013]: INFO nova.compute.manager [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Shelve offloading [ 1064.167724] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac6ed1f-238d-4c47-9fa0-917c6c39e471 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.176641] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d704174-ca12-414c-a034-cfae07aa5994 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.213094] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3547b92-e55f-4145-b9ce-1df16cb45aef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.216767] env[70013]: DEBUG nova.network.neutron [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "address": "fa:16:3e:2a:49:e7", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b19b577-68", "ovs_interfaceid": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84", "address": "fa:16:3e:a0:ee:70", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c185ee8-90", "ovs_interfaceid": "2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.225284] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b897b4ac-bb9b-4b91-a006-5d3db0695585 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.242669] env[70013]: DEBUG nova.compute.provider_tree [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.393309] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1064.393937] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e12e924a-e831-479b-bc89-8f376e131e67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.400923] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1064.400923] env[70013]: value = "task-4231321" [ 1064.400923] env[70013]: _type = "Task" [ 1064.400923] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.409461] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.509357] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.509720] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.719612] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1064.720346] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.720509] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.720810] env[70013]: DEBUG oslo_concurrency.lockutils [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.721007] env[70013]: DEBUG nova.network.neutron [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Refreshing network info cache for port 2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.723051] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792a3aab-f1b1-40d9-a3d4-653aa3a9ac81 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.750149] env[70013]: DEBUG nova.scheduler.client.report [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1064.754090] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1064.754358] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.754491] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1064.754681] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.754945] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1064.755052] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1064.755185] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1064.755351] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1064.755515] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1064.755678] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1064.755856] env[70013]: DEBUG nova.virt.hardware [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1064.762592] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfiguring VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1064.764016] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d56dba9a-9e1f-4e00-bbee-9fa6ba7c23fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.777264] env[70013]: DEBUG nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1064.787470] env[70013]: DEBUG oslo_vmware.api [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1064.787470] env[70013]: value = "task-4231322" [ 1064.787470] env[70013]: _type = "Task" [ 1064.787470] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.797631] env[70013]: DEBUG oslo_vmware.api [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231322, 'name': ReconfigVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.807037] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1064.807416] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.807679] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1064.807991] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.808266] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1064.808477] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1064.808904] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1064.809942] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1064.810143] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1064.810326] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1064.810502] env[70013]: DEBUG nova.virt.hardware [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1064.811470] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07df23ef-7d47-4071-b15c-e30ce846bcf0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.821204] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f44b93-4404-4aa5-9fd7-48bbf364d7c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.913012] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1064.913256] env[70013]: DEBUG nova.compute.manager [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1064.914074] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bcdd96-b1cd-40fc-8ec5-07eef3aa270f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.920676] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.920860] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.921034] env[70013]: DEBUG nova.network.neutron [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.012727] env[70013]: DEBUG nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1065.267027] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.267027] env[70013]: DEBUG nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1065.271141] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.229s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.271141] env[70013]: DEBUG nova.objects.instance [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lazy-loading 'resources' on Instance uuid 2c0e4326-c33a-42bb-b793-a100157b1c03 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.301086] env[70013]: DEBUG oslo_vmware.api [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231322, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.403782] env[70013]: DEBUG nova.compute.manager [req-8e28ee77-b409-49ce-85bd-f9ea84a1c517 req-15c49b05-37cd-487a-8993-06ba39be4ab4 service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Received event network-vif-plugged-f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1065.404079] env[70013]: DEBUG oslo_concurrency.lockutils [req-8e28ee77-b409-49ce-85bd-f9ea84a1c517 req-15c49b05-37cd-487a-8993-06ba39be4ab4 service nova] Acquiring lock "c57e876a-8696-4ce5-b533-7580f484414e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.404169] env[70013]: DEBUG oslo_concurrency.lockutils [req-8e28ee77-b409-49ce-85bd-f9ea84a1c517 req-15c49b05-37cd-487a-8993-06ba39be4ab4 service nova] Lock "c57e876a-8696-4ce5-b533-7580f484414e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.404339] env[70013]: DEBUG oslo_concurrency.lockutils [req-8e28ee77-b409-49ce-85bd-f9ea84a1c517 req-15c49b05-37cd-487a-8993-06ba39be4ab4 service nova] Lock "c57e876a-8696-4ce5-b533-7580f484414e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.404507] env[70013]: DEBUG nova.compute.manager [req-8e28ee77-b409-49ce-85bd-f9ea84a1c517 req-15c49b05-37cd-487a-8993-06ba39be4ab4 service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] No waiting events found dispatching network-vif-plugged-f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1065.404706] env[70013]: WARNING nova.compute.manager [req-8e28ee77-b409-49ce-85bd-f9ea84a1c517 req-15c49b05-37cd-487a-8993-06ba39be4ab4 service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Received unexpected event network-vif-plugged-f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 for instance with vm_state building and task_state spawning. [ 1065.491692] env[70013]: DEBUG nova.network.neutron [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Successfully updated port: f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.542852] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.561792] env[70013]: DEBUG nova.network.neutron [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updated VIF entry in instance network info cache for port 2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1065.562286] env[70013]: DEBUG nova.network.neutron [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "address": "fa:16:3e:2a:49:e7", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b19b577-68", "ovs_interfaceid": "7b19b577-689c-4e1d-bdb0-bab33f15bd8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84", "address": "fa:16:3e:a0:ee:70", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c185ee8-90", "ovs_interfaceid": "2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.773828] env[70013]: DEBUG nova.compute.utils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1065.776094] env[70013]: DEBUG nova.network.neutron [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fc25179-89", "ovs_interfaceid": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.777181] env[70013]: DEBUG nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1065.777380] env[70013]: DEBUG nova.network.neutron [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1065.799187] env[70013]: DEBUG oslo_vmware.api [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231322, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.824049] env[70013]: DEBUG nova.policy [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b9278c701c4426ab24082ed0a04744c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '940735ed506246c78bf5ede867e03751', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1065.996036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.996036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1065.996036] env[70013]: DEBUG nova.network.neutron [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.065047] env[70013]: DEBUG oslo_concurrency.lockutils [req-727c4429-5d49-4cdc-bb29-3c5cea61f31d req-82fd2614-d0fc-49a1-b7d3-2b4fc73e9db7 service nova] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.171519] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e340c2c3-db91-4036-a925-de6cafdc3064 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.179871] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb92024-d3f6-4cdf-b078-65c623174bb4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.215026] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1274d9f4-0849-4910-bcaa-f6d27ce9c2ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.224280] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec9cd09-fff9-4012-bee6-b9c09177c099 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.243320] env[70013]: DEBUG nova.compute.provider_tree [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.266852] env[70013]: DEBUG nova.network.neutron [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Successfully created port: 4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1066.283439] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.286464] env[70013]: DEBUG nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1066.299735] env[70013]: DEBUG oslo_vmware.api [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231322, 'name': ReconfigVM_Task, 'duration_secs': 1.28559} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.300160] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.300382] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfigured VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1066.543264] env[70013]: DEBUG nova.network.neutron [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1066.743439] env[70013]: DEBUG nova.network.neutron [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Updating instance_info_cache with network_info: [{"id": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "address": "fa:16:3e:a5:9a:1e", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf27a5efc-f9", "ovs_interfaceid": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.748019] env[70013]: DEBUG nova.scheduler.client.report [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1066.764456] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.765599] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd80c05-a687-43ba-b56c-bec6dda83133 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.775314] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1066.775621] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f8966f1-8569-408f-8507-3e524848286b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.805399] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c956513b-b1c2-4c61-a243-58436c9beb49 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.047s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.093965] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.094225] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.094417] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleting the datastore file [datastore2] 472b8218-51c3-492e-96ff-5fa99df4cc4c {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.094707] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f63bb2c9-14a0-4d97-b875-e88945efc464 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.103769] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1067.103769] env[70013]: value = "task-4231324" [ 1067.103769] env[70013]: _type = "Task" [ 1067.103769] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.113108] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.158890] env[70013]: INFO nova.compute.manager [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Rebuilding instance [ 1067.208755] env[70013]: DEBUG nova.compute.manager [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1067.209732] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9004a446-664e-47a2-a96e-3c9d104964c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.246414] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1067.246805] env[70013]: DEBUG nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Instance network_info: |[{"id": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "address": "fa:16:3e:a5:9a:1e", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf27a5efc-f9", "ovs_interfaceid": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1067.247333] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:9a:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f27a5efc-f9a7-4cb7-8db5-e1e692932fe1', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.255634] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Creating folder: Project (3c6094eeb5ca48dfbfbb10bc71feac67). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.257287] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.259578] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-687e4189-01b3-4749-9ab5-79524fb83a4b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.262459] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.556s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.262713] env[70013]: DEBUG oslo_concurrency.lockutils [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.263116] env[70013]: INFO nova.compute.manager [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] [instance: c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3] Successfully reverted task state from None on failure for instance. [ 1067.265658] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.909s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.265944] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.266104] env[70013]: INFO nova.compute.manager [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Successfully reverted task state from None on failure for instance. [ 1067.268611] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.336s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.268851] env[70013]: DEBUG nova.objects.instance [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lazy-loading 'resources' on Instance uuid 54d39dfd-6429-4009-8d26-22c1fa46243e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server [None req-ede2ff6b-8677-415c-8211-eebf504ae878 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-38840846-5423-4067-8d5a-02f68e981ac0"}]} [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1067.271406] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server [None req-abbc5eaa-1b24-4de2-a146-35702f406e35 tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1067.273446] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-92fcba01-94d0-4155-8c2b-93d8453ebad6"}]} [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1067.275392] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1067.277014] env[70013]: ERROR oslo_messaging.rpc.server [ 1067.282278] env[70013]: INFO nova.scheduler.client.report [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Deleted allocations for instance 2c0e4326-c33a-42bb-b793-a100157b1c03 [ 1067.284875] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Created folder: Project (3c6094eeb5ca48dfbfbb10bc71feac67) in parent group-v836999. [ 1067.285362] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Creating folder: Instances. Parent ref: group-v837195. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.287519] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2585f6e-51f8-4353-974d-ec598dd6ec5f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.295881] env[70013]: DEBUG nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1067.300731] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Created folder: Instances in parent group-v837195. [ 1067.301011] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1067.301239] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.302125] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d463de5e-8d1d-45af-bec5-211ffdf2c3b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.326723] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1067.327035] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1067.327201] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1067.327391] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1067.327541] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1067.327689] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1067.327929] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1067.328110] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1067.328293] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1067.328501] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1067.328624] env[70013]: DEBUG nova.virt.hardware [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1067.329595] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9659ad5f-d3ce-49ac-858d-1e46e5eb9fab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.333521] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.333521] env[70013]: value = "task-4231327" [ 1067.333521] env[70013]: _type = "Task" [ 1067.333521] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.342065] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26a3028-80bc-44d4-9822-822b07a90b65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.349683] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231327, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.515022] env[70013]: DEBUG nova.compute.manager [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Received event network-changed-f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1067.516411] env[70013]: DEBUG nova.compute.manager [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Refreshing instance network info cache due to event network-changed-f27a5efc-f9a7-4cb7-8db5-e1e692932fe1. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1067.516411] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Acquiring lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.516411] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Acquired lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.516411] env[70013]: DEBUG nova.network.neutron [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Refreshing network info cache for port f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.614906] env[70013]: DEBUG oslo_vmware.api [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170665} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.615268] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.615624] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.615887] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.636978] env[70013]: INFO nova.scheduler.client.report [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted allocations for instance 472b8218-51c3-492e-96ff-5fa99df4cc4c [ 1067.796051] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6b3a05b-e3f6-4105-9ffc-712bb611c89e tempest-AttachVolumeNegativeTest-242855104 tempest-AttachVolumeNegativeTest-242855104-project-member] Lock "2c0e4326-c33a-42bb-b793-a100157b1c03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.002s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.849600] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231327, 'name': CreateVM_Task, 'duration_secs': 0.423572} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.849778] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.850520] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.850775] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.851117] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1067.851407] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dacf2a6-eaba-428e-8338-be187617156e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.860282] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1067.860282] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52844d3e-4dce-f3e7-38e3-3a0774592e12" [ 1067.860282] env[70013]: _type = "Task" [ 1067.860282] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.872369] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52844d3e-4dce-f3e7-38e3-3a0774592e12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.137787] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e71c63-0e39-4eac-ba37-cc3d90337d30 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.142639] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.146850] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ce05c6-cd11-4a2e-8c41-0c1d7b5e18c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.188462] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436e087d-7912-4845-b481-e199e0dcbfa5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.198284] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ffae12-2e5a-4a77-9a15-a9d194605809 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.213991] env[70013]: DEBUG nova.compute.provider_tree [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.225670] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.225670] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c064f27d-9845-4e93-8119-c48e96542a60 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.234103] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1068.234103] env[70013]: value = "task-4231328" [ 1068.234103] env[70013]: _type = "Task" [ 1068.234103] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.243160] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.289823] env[70013]: DEBUG nova.network.neutron [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Updated VIF entry in instance network info cache for port f27a5efc-f9a7-4cb7-8db5-e1e692932fe1. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1068.290323] env[70013]: DEBUG nova.network.neutron [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Updating instance_info_cache with network_info: [{"id": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "address": "fa:16:3e:a5:9a:1e", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf27a5efc-f9", "ovs_interfaceid": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.317037] env[70013]: DEBUG nova.network.neutron [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Successfully updated port: 4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1068.348560] env[70013]: DEBUG nova.compute.manager [req-f7bfb30b-9b58-4402-8d4d-125375e32df4 req-059fe3b8-e672-4b21-aa9f-85bd9dc65e21 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Received event network-vif-plugged-4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1068.348560] env[70013]: DEBUG oslo_concurrency.lockutils [req-f7bfb30b-9b58-4402-8d4d-125375e32df4 req-059fe3b8-e672-4b21-aa9f-85bd9dc65e21 service nova] Acquiring lock "addf8ba9-142c-4458-8586-b06e52118d71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.348842] env[70013]: DEBUG oslo_concurrency.lockutils [req-f7bfb30b-9b58-4402-8d4d-125375e32df4 req-059fe3b8-e672-4b21-aa9f-85bd9dc65e21 service nova] Lock "addf8ba9-142c-4458-8586-b06e52118d71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.349145] env[70013]: DEBUG oslo_concurrency.lockutils [req-f7bfb30b-9b58-4402-8d4d-125375e32df4 req-059fe3b8-e672-4b21-aa9f-85bd9dc65e21 service nova] Lock "addf8ba9-142c-4458-8586-b06e52118d71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.349461] env[70013]: DEBUG nova.compute.manager [req-f7bfb30b-9b58-4402-8d4d-125375e32df4 req-059fe3b8-e672-4b21-aa9f-85bd9dc65e21 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] No waiting events found dispatching network-vif-plugged-4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1068.349767] env[70013]: WARNING nova.compute.manager [req-f7bfb30b-9b58-4402-8d4d-125375e32df4 req-059fe3b8-e672-4b21-aa9f-85bd9dc65e21 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Received unexpected event network-vif-plugged-4151335a-4b19-45bd-bf10-69e4960fe456 for instance with vm_state building and task_state spawning. [ 1068.375547] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52844d3e-4dce-f3e7-38e3-3a0774592e12, 'name': SearchDatastore_Task, 'duration_secs': 0.016832} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.375955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.376351] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.376671] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.376894] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.377195] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.377546] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e933a27-6d23-4eb9-b883-137de9593baf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.388398] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.388687] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1068.389840] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b367947-53e0-41c9-b26d-15e9ed905754 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.397502] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1068.397502] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c2e31b-6a7c-9044-301c-bc95d09aa124" [ 1068.397502] env[70013]: _type = "Task" [ 1068.397502] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.408283] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c2e31b-6a7c-9044-301c-bc95d09aa124, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.720771] env[70013]: DEBUG nova.scheduler.client.report [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1068.745516] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231328, 'name': PowerOffVM_Task, 'duration_secs': 0.217881} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.745785] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.746021] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.746816] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92aae08e-df28-4a62-9a34-d02b55bb5bfd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.755603] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.755866] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ec4804c-0d72-4394-b3ba-0bf8bdff524c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.796027] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Releasing lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.796027] env[70013]: DEBUG nova.compute.manager [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-vif-unplugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1068.796027] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.796027] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.796027] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.796027] env[70013]: DEBUG nova.compute.manager [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] No waiting events found dispatching network-vif-unplugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1068.796027] env[70013]: WARNING nova.compute.manager [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received unexpected event network-vif-unplugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 for instance with vm_state shelved and task_state shelving_offloading. [ 1068.796027] env[70013]: DEBUG nova.compute.manager [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1068.796027] env[70013]: DEBUG nova.compute.manager [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing instance network info cache due to event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1068.796027] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Acquiring lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.796027] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Acquired lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.796027] env[70013]: DEBUG nova.network.neutron [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1068.820064] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.820300] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquired lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.820383] env[70013]: DEBUG nova.network.neutron [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1068.839117] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.839480] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.839637] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleting the datastore file [datastore1] 1230ca35-1512-4464-85fc-d3b4ab05eac1 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.840241] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6ea7a40-a70f-4c6b-85b4-fe49b554779e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.848278] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1068.848278] env[70013]: value = "task-4231330" [ 1068.848278] env[70013]: _type = "Task" [ 1068.848278] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.857962] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.909025] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c2e31b-6a7c-9044-301c-bc95d09aa124, 'name': SearchDatastore_Task, 'duration_secs': 0.010556} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.909842] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-412cdfa9-df91-44ef-9a1d-d204e81046c2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.916048] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1068.916048] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528906e4-6f50-2440-ad06-3058c4c6081e" [ 1068.916048] env[70013]: _type = "Task" [ 1068.916048] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.925056] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528906e4-6f50-2440-ad06-3058c4c6081e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.227248] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.229812] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.177s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.230021] env[70013]: DEBUG oslo_concurrency.lockutils [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.230280] env[70013]: INFO nova.compute.manager [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Successfully reverted task state from None on failure for instance. [ 1069.232807] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.026s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.234362] env[70013]: INFO nova.compute.claims [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server [None req-87c5bfe3-23ea-4bf6-a819-b1c48dbebe64 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1406, in catch_all [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server yield [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1011, in set_inventory_for_provider [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/dc0a4c9d-b611-453e-a900-41e280c29b95/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'type': 'integer', 'maximum': 2147483647, 'minimum': 1} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c070cdc0-931e-46e4-9c1f-fe292e9e6ad8"}]} [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 167, in decorated_function [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 158, in decorated_function [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 214, in decorated_function [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 204, in decorated_function [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3388, in terminate_instance [ 1069.240135] env[70013]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3383, in do_terminate_instance [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server raise self.value [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3376, in do_terminate_instance [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3340, in _delete_instance [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 929, in _complete_deletion [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 695, in _update_resource_tracker [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 732, in update_usage [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1405, in _update [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 724, in reraise [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server raise value [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1390, in _update_to_placement [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1501, in update_from_provider_tree [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1418, in catch_all [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 1069.241566] env[70013]: ERROR oslo_messaging.rpc.server [ 1069.250299] env[70013]: INFO nova.scheduler.client.report [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Deleted allocations for instance 54d39dfd-6429-4009-8d26-22c1fa46243e [ 1069.365489] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148737} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.366408] env[70013]: DEBUG nova.network.neutron [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1069.368308] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.368499] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.368677] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.428580] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528906e4-6f50-2440-ad06-3058c4c6081e, 'name': SearchDatastore_Task, 'duration_secs': 0.010011} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.428848] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1069.429123] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/c57e876a-8696-4ce5-b533-7580f484414e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1069.429427] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f501c213-340d-418d-9278-aeadb67ea104 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.437519] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1069.437519] env[70013]: value = "task-4231331" [ 1069.437519] env[70013]: _type = "Task" [ 1069.437519] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.450081] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231331, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.474576] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-7b19b577-689c-4e1d-bdb0-bab33f15bd8f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.474881] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-7b19b577-689c-4e1d-bdb0-bab33f15bd8f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.521895] env[70013]: DEBUG nova.network.neutron [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Updating instance_info_cache with network_info: [{"id": "4151335a-4b19-45bd-bf10-69e4960fe456", "address": "fa:16:3e:2c:5d:43", "network": {"id": "a6808f1c-7ccf-4a81-9bac-cdbab50a0a2a", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1787810678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940735ed506246c78bf5ede867e03751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4151335a-4b", "ovs_interfaceid": "4151335a-4b19-45bd-bf10-69e4960fe456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.757350] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b25ce846-f889-4ef6-a8f6-a42c287dcbca tempest-ListServerFiltersTestJSON-2066587110 tempest-ListServerFiltersTestJSON-2066587110-project-member] Lock "54d39dfd-6429-4009-8d26-22c1fa46243e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.382s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.861582] env[70013]: DEBUG nova.network.neutron [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updated VIF entry in instance network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1069.862054] env[70013]: DEBUG nova.network.neutron [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap3fc25179-89", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.949668] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231331, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.982034] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.982261] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1069.982706] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.983932] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c904468a-f356-4596-9c6d-436d6f3a95f4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.003101] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b293645e-be58-4a51-a1e9-39fe69ec02cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.034681] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Releasing lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.035091] env[70013]: DEBUG nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Instance network_info: |[{"id": "4151335a-4b19-45bd-bf10-69e4960fe456", "address": "fa:16:3e:2c:5d:43", "network": {"id": "a6808f1c-7ccf-4a81-9bac-cdbab50a0a2a", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1787810678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940735ed506246c78bf5ede867e03751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4151335a-4b", "ovs_interfaceid": "4151335a-4b19-45bd-bf10-69e4960fe456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1070.041652] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfiguring VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1070.042135] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:5d:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '415e68b4-3766-4359-afe2-f8563910d98c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4151335a-4b19-45bd-bf10-69e4960fe456', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.051612] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Creating folder: Project (940735ed506246c78bf5ede867e03751). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1070.052104] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0edc14e9-b281-44d1-b69b-0cc2083f4016 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.065480] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-503e53ce-59a4-449f-a866-2d99b64b7f0e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.075075] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1070.075075] env[70013]: value = "task-4231333" [ 1070.075075] env[70013]: _type = "Task" [ 1070.075075] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.079710] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Created folder: Project (940735ed506246c78bf5ede867e03751) in parent group-v836999. [ 1070.079921] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Creating folder: Instances. Parent ref: group-v837198. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1070.080626] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7fa9834b-210b-4e0b-b587-2892b81109e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.085911] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.090893] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Created folder: Instances in parent group-v837198. [ 1070.091160] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1070.091457] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1070.091664] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2aa3e676-0b03-447f-b93f-ab155262e4a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.111230] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.111230] env[70013]: value = "task-4231335" [ 1070.111230] env[70013]: _type = "Task" [ 1070.111230] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.120042] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231335, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.368178] env[70013]: DEBUG oslo_concurrency.lockutils [req-c4acc7d7-00ce-4b29-ab29-6ef0612c5de1 req-4613820b-d265-4ccb-984d-2e54e2352d6b service nova] Releasing lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.389019] env[70013]: DEBUG nova.compute.manager [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Received event network-changed-4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1070.389019] env[70013]: DEBUG nova.compute.manager [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Refreshing instance network info cache due to event network-changed-4151335a-4b19-45bd-bf10-69e4960fe456. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1070.389019] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] Acquiring lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.389019] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] Acquired lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1070.389019] env[70013]: DEBUG nova.network.neutron [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Refreshing network info cache for port 4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1070.409137] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1070.409137] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.409294] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1070.409480] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.409632] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1070.409780] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1070.409989] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1070.410283] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1070.410473] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1070.410656] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1070.410834] env[70013]: DEBUG nova.virt.hardware [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1070.412304] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adabaae-1e37-4f64-b193-c33a75126b90 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.424243] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eaef595-db1b-4c31-9767-0e94469b0665 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.441028] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:e4:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'abcf0d10-3f3f-45dc-923e-1c78766e2dad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d1c82dc-c975-4156-8d68-368f68ec3f87', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.448788] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1070.454207] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1070.454739] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecc10ef8-df59-4f20-9639-463ac7d97b89 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.478396] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231331, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.479925] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.479925] env[70013]: value = "task-4231336" [ 1070.479925] env[70013]: _type = "Task" [ 1070.479925] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.489983] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231336, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.587799] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.621764] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231335, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.647212] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52df952-bbf6-4b33-a4fc-33423cd753fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.654681] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39622129-6c2b-430c-9891-0312160bfbe3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.693010] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb9a4b4-917a-43c6-bcb2-78b999c6baa3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.701955] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5544df3-3969-46a6-8adb-521c1ce321df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.718190] env[70013]: DEBUG nova.compute.provider_tree [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.953049] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231331, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.991903] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231336, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.090766] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.123565] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231335, 'name': CreateVM_Task, 'duration_secs': 0.983815} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.123780] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.124524] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.124722] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.125085] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1071.125356] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e08edd7a-07d1-4138-b2e2-a1ec5528fd8e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.131156] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1071.131156] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dfb014-20e2-7512-bdce-ba05961db411" [ 1071.131156] env[70013]: _type = "Task" [ 1071.131156] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.140437] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dfb014-20e2-7512-bdce-ba05961db411, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.222450] env[70013]: DEBUG nova.scheduler.client.report [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1071.295549] env[70013]: DEBUG nova.network.neutron [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Updated VIF entry in instance network info cache for port 4151335a-4b19-45bd-bf10-69e4960fe456. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1071.295988] env[70013]: DEBUG nova.network.neutron [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Updating instance_info_cache with network_info: [{"id": "4151335a-4b19-45bd-bf10-69e4960fe456", "address": "fa:16:3e:2c:5d:43", "network": {"id": "a6808f1c-7ccf-4a81-9bac-cdbab50a0a2a", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1787810678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940735ed506246c78bf5ede867e03751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4151335a-4b", "ovs_interfaceid": "4151335a-4b19-45bd-bf10-69e4960fe456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.452877] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231331, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.5421} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.453203] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/c57e876a-8696-4ce5-b533-7580f484414e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1071.453440] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.453728] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9fd3e170-bdfd-4197-8790-af4cbca19d4d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.461783] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1071.461783] env[70013]: value = "task-4231337" [ 1071.461783] env[70013]: _type = "Task" [ 1071.461783] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.471397] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231337, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.489990] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231336, 'name': CreateVM_Task, 'duration_secs': 0.840069} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.490215] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.490956] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.591272] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.646590] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dfb014-20e2-7512-bdce-ba05961db411, 'name': SearchDatastore_Task, 'duration_secs': 0.010823} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.646944] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1071.647208] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.647453] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.647602] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.647783] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.648125] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.648504] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1071.648754] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed303d71-e109-4bba-9af2-6b97cdcc948f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.653193] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7804c3b-7a91-4da0-88f9-a52a35af983a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.661376] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1071.661376] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca4e04-e043-d898-275e-9fa57afb09a8" [ 1071.661376] env[70013]: _type = "Task" [ 1071.661376] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.667671] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.668188] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1071.670086] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-644249e8-3b68-4e09-befb-67a86e51535c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.678212] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ca4e04-e043-d898-275e-9fa57afb09a8, 'name': SearchDatastore_Task, 'duration_secs': 0.010687} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.679213] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1071.679605] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.680222] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.683072] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1071.683072] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52831fc8-65e4-cdff-181f-c0e71e103d80" [ 1071.683072] env[70013]: _type = "Task" [ 1071.683072] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.692502] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52831fc8-65e4-cdff-181f-c0e71e103d80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.728798] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1071.729412] env[70013]: DEBUG nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1071.732333] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.254s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.732691] env[70013]: DEBUG nova.objects.instance [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'resources' on Instance uuid 2070989d-6c90-4eb1-8508-7587ed9659d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.799895] env[70013]: DEBUG oslo_concurrency.lockutils [req-0a7de1a2-dc7f-4c36-b598-c74f74c7772b req-52da5a05-d50b-4484-a528-35e7e8002a54 service nova] Releasing lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1071.972194] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231337, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075788} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.972583] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.973445] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e57373-e50e-4650-8c97-e669b620a640 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.997489] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/c57e876a-8696-4ce5-b533-7580f484414e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.997800] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ebdd626-c014-4212-a223-407410f675ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.018515] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1072.018515] env[70013]: value = "task-4231338" [ 1072.018515] env[70013]: _type = "Task" [ 1072.018515] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.031879] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231338, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.095390] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.194676] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52831fc8-65e4-cdff-181f-c0e71e103d80, 'name': SearchDatastore_Task, 'duration_secs': 0.010049} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.195809] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16cc402f-3347-433e-b074-76ae9d2e89e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.210696] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1072.210696] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c78831-0a5f-24b0-3056-4a9884f98a1b" [ 1072.210696] env[70013]: _type = "Task" [ 1072.210696] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.223315] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c78831-0a5f-24b0-3056-4a9884f98a1b, 'name': SearchDatastore_Task, 'duration_secs': 0.011013} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.224253] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1072.224519] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] addf8ba9-142c-4458-8586-b06e52118d71/addf8ba9-142c-4458-8586-b06e52118d71.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1072.225048] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1072.225288] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.225517] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f98c5069-62a0-43c6-b6c1-1121753acacc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.227883] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be4358dd-7789-4b5c-a393-e9d35eaf91fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.235989] env[70013]: DEBUG nova.compute.utils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1072.241490] env[70013]: DEBUG nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1072.241960] env[70013]: DEBUG nova.network.neutron [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1072.244547] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1072.244547] env[70013]: value = "task-4231339" [ 1072.244547] env[70013]: _type = "Task" [ 1072.244547] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.246521] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.246714] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1072.251752] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-921f72e8-8c3b-49ee-82e0-6a8d66aeab78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.259104] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1072.259104] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52619c52-a6da-661d-40da-296c84beaf1d" [ 1072.259104] env[70013]: _type = "Task" [ 1072.259104] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.263636] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.277991] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52619c52-a6da-661d-40da-296c84beaf1d, 'name': SearchDatastore_Task, 'duration_secs': 0.012259} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.280402] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46d7acbd-a92c-4364-86a6-6261b5d00690 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.288678] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1072.288678] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527397c6-cf21-0d77-d034-82229f28a636" [ 1072.288678] env[70013]: _type = "Task" [ 1072.288678] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.296708] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527397c6-cf21-0d77-d034-82229f28a636, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.322669] env[70013]: DEBUG nova.policy [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c4feb0b50fc4d83b93d6b9290848ee3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '946b799447b74a32b252c4f69fa900cd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1072.534801] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231338, 'name': ReconfigVM_Task, 'duration_secs': 0.307193} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.534801] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Reconfigured VM instance instance-0000005d to attach disk [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/c57e876a-8696-4ce5-b533-7580f484414e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.535832] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee402f96-4953-4086-9df6-ccd1f7e9c1c3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.549088] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1072.549088] env[70013]: value = "task-4231340" [ 1072.549088] env[70013]: _type = "Task" [ 1072.549088] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.567677] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231340, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.601086] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.702147] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79984be8-764e-4a72-8e16-67455649aa8f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.707524] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.707953] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.708132] env[70013]: INFO nova.compute.manager [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Shelving [ 1072.713462] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8f3b13-8227-4196-a58c-6a7fcbb71644 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.762674] env[70013]: DEBUG nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1072.773679] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a88fbb8-c2cc-406b-a8f5-835ba3cc3c56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.788023] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231339, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.789035] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7165b2e-b48c-4e94-a286-80dff8b76430 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.808848] env[70013]: DEBUG nova.compute.provider_tree [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.814199] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527397c6-cf21-0d77-d034-82229f28a636, 'name': SearchDatastore_Task, 'duration_secs': 0.021029} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.814745] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1072.815769] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1072.815914] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdf7cfdf-162f-47e3-ad4c-f6f7ae7d9498 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.824563] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1072.824563] env[70013]: value = "task-4231341" [ 1072.824563] env[70013]: _type = "Task" [ 1072.824563] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.831262] env[70013]: DEBUG nova.network.neutron [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Successfully created port: c617a733-3882-4dd2-842d-f2be129d99ae {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1072.841331] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.061048] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231340, 'name': Rename_Task, 'duration_secs': 0.305992} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.061048] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.061768] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f5a0751-ac30-45ea-a40a-26a37d998657 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.071322] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1073.071322] env[70013]: value = "task-4231342" [ 1073.071322] env[70013]: _type = "Task" [ 1073.071322] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.089771] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231342, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.101625] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.285895] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547132} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.285895] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] addf8ba9-142c-4458-8586-b06e52118d71/addf8ba9-142c-4458-8586-b06e52118d71.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1073.285895] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.285895] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f588d522-7462-4ce1-932c-97bd30ff39c2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.294835] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1073.294835] env[70013]: value = "task-4231343" [ 1073.294835] env[70013]: _type = "Task" [ 1073.294835] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.307533] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231343, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.315943] env[70013]: DEBUG nova.scheduler.client.report [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1073.335723] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497594} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.336028] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1073.336257] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.336522] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f925cfc1-9f0a-4144-97a9-87124b9ab933 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.344518] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1073.344518] env[70013]: value = "task-4231344" [ 1073.344518] env[70013]: _type = "Task" [ 1073.344518] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.354244] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231344, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.582034] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231342, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.598748] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.727572] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.727572] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f23b9a63-005d-4262-94fb-7524eb53d4b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.734062] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1073.734062] env[70013]: value = "task-4231345" [ 1073.734062] env[70013]: _type = "Task" [ 1073.734062] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.744681] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231345, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.778046] env[70013]: DEBUG nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1073.809232] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231343, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090804} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.810511] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1073.810817] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.811047] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1073.811470] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.811553] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1073.811735] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1073.811998] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1073.812196] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1073.812405] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1073.812568] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1073.812822] env[70013]: DEBUG nova.virt.hardware [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1073.813192] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.814412] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930c8e3a-ff74-431a-adea-e9f2dc2ccbeb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.818333] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a643a3-069b-4d82-a755-d5a23abcf684 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.824103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.092s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1073.839652] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.026s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1073.851497] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] addf8ba9-142c-4458-8586-b06e52118d71/addf8ba9-142c-4458-8586-b06e52118d71.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.852822] env[70013]: INFO nova.scheduler.client.report [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted allocations for instance 2070989d-6c90-4eb1-8508-7587ed9659d6 [ 1073.855029] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c034d33-1210-4ff9-a61d-05d5770ef38c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.859745] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d90e744b-9f50-4ae1-8b70-adc5458917ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.896906] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1073.896906] env[70013]: value = "task-4231346" [ 1073.896906] env[70013]: _type = "Task" [ 1073.896906] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.897292] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231344, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072133} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.897733] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.901916] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c707d8b-8c02-49e9-9d46-8f870198b5fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.934286] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.934710] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231346, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.937158] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5f27141-cc6f-41af-8a32-1559954f8975 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.960314] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1073.960314] env[70013]: value = "task-4231347" [ 1073.960314] env[70013]: _type = "Task" [ 1073.960314] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.970309] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231347, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.082085] env[70013]: DEBUG oslo_vmware.api [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231342, 'name': PowerOnVM_Task, 'duration_secs': 0.576209} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.082496] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.082748] env[70013]: INFO nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Took 9.31 seconds to spawn the instance on the hypervisor. [ 1074.084752] env[70013]: DEBUG nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1074.084752] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a42612-b82d-436e-ab25-6193576b7428 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.098659] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.172833] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "be86d96e-15be-4c39-b05d-59da0668c0d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.173095] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.244309] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231345, 'name': PowerOffVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.387947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-a7b7b4c3-7ee2-4300-a80d-15b91eb7673c tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "2070989d-6c90-4eb1-8508-7587ed9659d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.432s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.412853] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231346, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.474390] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231347, 'name': ReconfigVM_Task, 'duration_secs': 0.286381} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.474390] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 1230ca35-1512-4464-85fc-d3b4ab05eac1/1230ca35-1512-4464-85fc-d3b4ab05eac1.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.474686] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79b36c5b-d46c-4897-956a-b020f4dd1370 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.482640] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1074.482640] env[70013]: value = "task-4231348" [ 1074.482640] env[70013]: _type = "Task" [ 1074.482640] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.501547] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231348, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.513062] env[70013]: DEBUG nova.compute.manager [req-3b7a5624-0e54-4ac0-bf94-1c07e38f8b8e req-13b59590-ae4b-4dd3-aeb4-39b0dbbd7dc3 service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Received event network-vif-plugged-c617a733-3882-4dd2-842d-f2be129d99ae {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1074.513062] env[70013]: DEBUG oslo_concurrency.lockutils [req-3b7a5624-0e54-4ac0-bf94-1c07e38f8b8e req-13b59590-ae4b-4dd3-aeb4-39b0dbbd7dc3 service nova] Acquiring lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.513062] env[70013]: DEBUG oslo_concurrency.lockutils [req-3b7a5624-0e54-4ac0-bf94-1c07e38f8b8e req-13b59590-ae4b-4dd3-aeb4-39b0dbbd7dc3 service nova] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.513062] env[70013]: DEBUG oslo_concurrency.lockutils [req-3b7a5624-0e54-4ac0-bf94-1c07e38f8b8e req-13b59590-ae4b-4dd3-aeb4-39b0dbbd7dc3 service nova] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.513062] env[70013]: DEBUG nova.compute.manager [req-3b7a5624-0e54-4ac0-bf94-1c07e38f8b8e req-13b59590-ae4b-4dd3-aeb4-39b0dbbd7dc3 service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] No waiting events found dispatching network-vif-plugged-c617a733-3882-4dd2-842d-f2be129d99ae {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1074.513399] env[70013]: WARNING nova.compute.manager [req-3b7a5624-0e54-4ac0-bf94-1c07e38f8b8e req-13b59590-ae4b-4dd3-aeb4-39b0dbbd7dc3 service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Received unexpected event network-vif-plugged-c617a733-3882-4dd2-842d-f2be129d99ae for instance with vm_state building and task_state spawning. [ 1074.598822] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.615895] env[70013]: INFO nova.compute.manager [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Took 31.53 seconds to build instance. [ 1074.673314] env[70013]: DEBUG nova.network.neutron [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Successfully updated port: c617a733-3882-4dd2-842d-f2be129d99ae {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1074.680091] env[70013]: DEBUG nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1074.746904] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231345, 'name': PowerOffVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.844334] env[70013]: INFO nova.compute.manager [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Rescuing [ 1074.844623] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.844776] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1074.844944] env[70013]: DEBUG nova.network.neutron [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1074.910128] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231346, 'name': ReconfigVM_Task, 'duration_secs': 0.856478} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.910512] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Reconfigured VM instance instance-0000005e to attach disk [datastore2] addf8ba9-142c-4458-8586-b06e52118d71/addf8ba9-142c-4458-8586-b06e52118d71.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.911529] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00364d9b-5e87-410a-a8a2-0a2c3ba31dc3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.919718] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1074.919718] env[70013]: value = "task-4231349" [ 1074.919718] env[70013]: _type = "Task" [ 1074.919718] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.929984] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231349, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.993820] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231348, 'name': Rename_Task, 'duration_secs': 0.139738} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.994246] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.994714] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af1de781-159a-4d3c-be12-a630ca0177c1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.003031] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1075.003031] env[70013]: value = "task-4231350" [ 1075.003031] env[70013]: _type = "Task" [ 1075.003031] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.016415] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231350, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.101758] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.118609] env[70013]: DEBUG oslo_concurrency.lockutils [None req-015eab31-1690-4179-ac7b-4bcc59abb595 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "c57e876a-8696-4ce5-b533-7580f484414e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.042s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.175980] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "refresh_cache-930ababf-5f2c-4b7e-a24d-7831fbd73d66" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.176230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "refresh_cache-930ababf-5f2c-4b7e-a24d-7831fbd73d66" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.176427] env[70013]: DEBUG nova.network.neutron [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1075.206522] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.246049] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231345, 'name': PowerOffVM_Task, 'duration_secs': 1.223704} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.246409] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.247492] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2beb2317-0f13-457b-8730-fc37d6bed584 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.286269] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50954d3-4b57-44f4-8056-8a8f2b55a8a3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.430778] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231349, 'name': Rename_Task, 'duration_secs': 0.318798} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.433428] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.433704] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-222abbcf-6325-48cc-81e5-7811ddbc1afe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.441053] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1075.441053] env[70013]: value = "task-4231351" [ 1075.441053] env[70013]: _type = "Task" [ 1075.441053] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.449191] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231351, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.515235] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231350, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.569145] env[70013]: DEBUG nova.network.neutron [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Updating instance_info_cache with network_info: [{"id": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "address": "fa:16:3e:a5:9a:1e", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf27a5efc-f9", "ovs_interfaceid": "f27a5efc-f9a7-4cb7-8db5-e1e692932fe1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.601257] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.715422] env[70013]: DEBUG nova.network.neutron [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1075.757198] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "e3b37310-262a-4852-89b4-4d1b1d6f5776" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.757453] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.799995] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1075.800381] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8b9527ef-e66c-49b6-a7c4-a520711bdd84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.808470] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1075.808470] env[70013]: value = "task-4231352" [ 1075.808470] env[70013]: _type = "Task" [ 1075.808470] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.819465] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231352, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.953535] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231351, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.013933] env[70013]: DEBUG oslo_vmware.api [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231350, 'name': PowerOnVM_Task, 'duration_secs': 0.655483} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.014282] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.014511] env[70013]: DEBUG nova.compute.manager [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1076.015409] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d69d68-e1b2-4392-8530-22a8514aea19 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.072552] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "refresh_cache-c57e876a-8696-4ce5-b533-7580f484414e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1076.105814] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.190525] env[70013]: DEBUG nova.network.neutron [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Updating instance_info_cache with network_info: [{"id": "c617a733-3882-4dd2-842d-f2be129d99ae", "address": "fa:16:3e:8d:59:ac", "network": {"id": "0f803fec-8c1a-495e-b561-9ae7e3b7c4e2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-731158857-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "946b799447b74a32b252c4f69fa900cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc617a733-38", "ovs_interfaceid": "c617a733-3882-4dd2-842d-f2be129d99ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.260900] env[70013]: DEBUG nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1076.321761] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231352, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.453068] env[70013]: DEBUG oslo_vmware.api [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231351, 'name': PowerOnVM_Task, 'duration_secs': 0.685848} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.453333] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.453563] env[70013]: INFO nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Took 9.16 seconds to spawn the instance on the hypervisor. [ 1076.453778] env[70013]: DEBUG nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1076.454663] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16444db4-7530-4fd3-b80e-243ad7c85210 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.534329] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.541702] env[70013]: DEBUG nova.compute.manager [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Received event network-changed-c617a733-3882-4dd2-842d-f2be129d99ae {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1076.541881] env[70013]: DEBUG nova.compute.manager [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Refreshing instance network info cache due to event network-changed-c617a733-3882-4dd2-842d-f2be129d99ae. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1076.542157] env[70013]: DEBUG oslo_concurrency.lockutils [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] Acquiring lock "refresh_cache-930ababf-5f2c-4b7e-a24d-7831fbd73d66" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.602180] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.693883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "refresh_cache-930ababf-5f2c-4b7e-a24d-7831fbd73d66" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1076.694278] env[70013]: DEBUG nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Instance network_info: |[{"id": "c617a733-3882-4dd2-842d-f2be129d99ae", "address": "fa:16:3e:8d:59:ac", "network": {"id": "0f803fec-8c1a-495e-b561-9ae7e3b7c4e2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-731158857-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "946b799447b74a32b252c4f69fa900cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc617a733-38", "ovs_interfaceid": "c617a733-3882-4dd2-842d-f2be129d99ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1076.694619] env[70013]: DEBUG oslo_concurrency.lockutils [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] Acquired lock "refresh_cache-930ababf-5f2c-4b7e-a24d-7831fbd73d66" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.694800] env[70013]: DEBUG nova.network.neutron [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Refreshing network info cache for port c617a733-3882-4dd2-842d-f2be129d99ae {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1076.696986] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:59:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a55f45a-d631-4ebc-b73b-8a30bd0a32a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c617a733-3882-4dd2-842d-f2be129d99ae', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.705536] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1076.707159] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.707159] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fcd1573b-b7e2-4e43-9412-fb451aa7a8d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.730008] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.730008] env[70013]: value = "task-4231353" [ 1076.730008] env[70013]: _type = "Task" [ 1076.730008] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.739146] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231353, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.786051] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.821136] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231352, 'name': CreateSnapshot_Task, 'duration_secs': 0.804846} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.821477] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1076.822307] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149adf1b-b7f5-4004-95e0-ff665ac2522a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.973959] env[70013]: INFO nova.compute.manager [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Took 28.84 seconds to build instance. [ 1077.103635] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.245029] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231353, 'name': CreateVM_Task, 'duration_secs': 0.413142} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.245413] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.246353] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.246641] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.247155] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1077.247536] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cc4ab96-3e8a-4e46-91e9-8fe28e9c161d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.255096] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1077.255096] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a929e7-302c-1ee8-939a-168d41829c71" [ 1077.255096] env[70013]: _type = "Task" [ 1077.255096] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.267770] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a929e7-302c-1ee8-939a-168d41829c71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.344174] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1077.344531] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c580670c-f4e8-47e0-b735-15f748f996cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.354303] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1077.354303] env[70013]: value = "task-4231354" [ 1077.354303] env[70013]: _type = "Task" [ 1077.354303] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.365730] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231354, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.477809] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5ac78b0-1ac5-447b-b57c-39843d397fed tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "addf8ba9-142c-4458-8586-b06e52118d71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.369s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1077.601663] env[70013]: DEBUG oslo_vmware.api [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231333, 'name': ReconfigVM_Task, 'duration_secs': 7.199658} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.601915] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.602140] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Reconfigured VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1077.618111] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1077.618448] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f39b539b-fa59-4eb8-816b-efda121c58cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.627946] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1077.627946] env[70013]: value = "task-4231355" [ 1077.627946] env[70013]: _type = "Task" [ 1077.627946] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.639998] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.766996] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a929e7-302c-1ee8-939a-168d41829c71, 'name': SearchDatastore_Task, 'duration_secs': 0.037581} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.767911] env[70013]: DEBUG nova.network.neutron [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Updated VIF entry in instance network info cache for port c617a733-3882-4dd2-842d-f2be129d99ae. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1077.768257] env[70013]: DEBUG nova.network.neutron [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Updating instance_info_cache with network_info: [{"id": "c617a733-3882-4dd2-842d-f2be129d99ae", "address": "fa:16:3e:8d:59:ac", "network": {"id": "0f803fec-8c1a-495e-b561-9ae7e3b7c4e2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-731158857-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "946b799447b74a32b252c4f69fa900cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc617a733-38", "ovs_interfaceid": "c617a733-3882-4dd2-842d-f2be129d99ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.769549] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.769774] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.770126] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.770374] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.770671] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.771328] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a831528e-930c-44fc-b2f7-695fdb09c83c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.782154] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.782599] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1077.783958] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bb48418-04e6-4cf7-896f-ea0ea1ba24ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.791962] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1077.791962] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf495e-389d-26d7-47c7-3843d459606d" [ 1077.791962] env[70013]: _type = "Task" [ 1077.791962] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.804565] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf495e-389d-26d7-47c7-3843d459606d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.866490] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231354, 'name': CloneVM_Task} progress is 93%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.915574] env[70013]: DEBUG nova.compute.manager [req-98505f27-a0d8-4ee5-a8c0-c18c6b535abc req-77198630-e4b4-43c5-b836-e818346330d5 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-vif-deleted-7b19b577-689c-4e1d-bdb0-bab33f15bd8f {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1077.916642] env[70013]: INFO nova.compute.manager [req-98505f27-a0d8-4ee5-a8c0-c18c6b535abc req-77198630-e4b4-43c5-b836-e818346330d5 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Neutron deleted interface 7b19b577-689c-4e1d-bdb0-bab33f15bd8f; detaching it from the instance and deleting it from the info cache [ 1077.916642] env[70013]: DEBUG nova.network.neutron [req-98505f27-a0d8-4ee5-a8c0-c18c6b535abc req-77198630-e4b4-43c5-b836-e818346330d5 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84", "address": "fa:16:3e:a0:ee:70", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c185ee8-90", "ovs_interfaceid": "2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.140884] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231355, 'name': PowerOffVM_Task, 'duration_secs': 0.393653} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.141224] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1078.142198] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ffbae2-1520-43b3-bf91-8a2c3f8df22e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.164090] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdcf580-1904-4f21-b5e9-8349488ccec2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.199772] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.200094] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6486efbe-7e6b-41ad-a896-03d2c79623b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.207898] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1078.207898] env[70013]: value = "task-4231356" [ 1078.207898] env[70013]: _type = "Task" [ 1078.207898] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.218401] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231356, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.272748] env[70013]: DEBUG oslo_concurrency.lockutils [req-8a35682d-a955-433c-8db2-111ab7254e24 req-afe78029-b006-495d-8be2-4a5b2bd6998a service nova] Releasing lock "refresh_cache-930ababf-5f2c-4b7e-a24d-7831fbd73d66" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.304885] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf495e-389d-26d7-47c7-3843d459606d, 'name': SearchDatastore_Task, 'duration_secs': 0.015119} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.305858] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ac7ed56-0bfa-4481-aa4d-4f8e07486fe6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.313547] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1078.313547] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5286222e-b1b5-4458-5650-c75fc4e1ee72" [ 1078.313547] env[70013]: _type = "Task" [ 1078.313547] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.325736] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5286222e-b1b5-4458-5650-c75fc4e1ee72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.367511] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231354, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.419966] env[70013]: DEBUG oslo_concurrency.lockutils [req-98505f27-a0d8-4ee5-a8c0-c18c6b535abc req-77198630-e4b4-43c5-b836-e818346330d5 service nova] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.420119] env[70013]: DEBUG oslo_concurrency.lockutils [req-98505f27-a0d8-4ee5-a8c0-c18c6b535abc req-77198630-e4b4-43c5-b836-e818346330d5 service nova] Acquired lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.421183] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70266a0f-fa66-423e-b941-754348166e8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.425619] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 28a56a75-9b30-4121-8252-a9e57287441c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.425806] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance fc473d40-b57b-437e-9511-58a0ba700a69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.425954] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.426128] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 09544922-6dc4-48b5-8cfd-e91e7f74c13f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.426280] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.426424] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.426606] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.426779] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.426910] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.427079] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.427261] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 0ea84d0d-905c-428d-8abb-2781c817f08f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.427454] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.427601] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.427749] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 1230ca35-1512-4464-85fc-d3b4ab05eac1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.427937] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 181238eb-cb0c-4740-9896-ca745b53ebe8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1078.428105] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9026841c-7bda-41e5-a4ac-03d0d3e37560 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.428264] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.428415] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8810bebc-448c-4e9b-9fbb-ce9e24611dc9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.428587] env[70013]: WARNING nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ec7d0e9a-4a73-4681-9395-a3c963772f62 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1078.428731] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c57e876a-8696-4ce5-b533-7580f484414e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.428881] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance addf8ba9-142c-4458-8586-b06e52118d71 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.429074] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 930ababf-5f2c-4b7e-a24d-7831fbd73d66 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1078.447264] env[70013]: DEBUG oslo_concurrency.lockutils [req-98505f27-a0d8-4ee5-a8c0-c18c6b535abc req-77198630-e4b4-43c5-b836-e818346330d5 service nova] Releasing lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.447558] env[70013]: WARNING nova.compute.manager [req-98505f27-a0d8-4ee5-a8c0-c18c6b535abc req-77198630-e4b4-43c5-b836-e818346330d5 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Detach interface failed, port_id=7b19b577-689c-4e1d-bdb0-bab33f15bd8f, reason: No device with interface-id 7b19b577-689c-4e1d-bdb0-bab33f15bd8f exists on VM: nova.exception.NotFound: No device with interface-id 7b19b577-689c-4e1d-bdb0-bab33f15bd8f exists on VM [ 1078.705837] env[70013]: DEBUG nova.compute.manager [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Received event network-changed-4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1078.705837] env[70013]: DEBUG nova.compute.manager [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Refreshing instance network info cache due to event network-changed-4151335a-4b19-45bd-bf10-69e4960fe456. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1078.706157] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] Acquiring lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.706997] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] Acquired lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.707140] env[70013]: DEBUG nova.network.neutron [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Refreshing network info cache for port 4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1078.725817] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1078.726165] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.726465] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.726682] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.726941] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.727522] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71a2bcb7-dad0-49c7-8615-ffa6209ce569 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.740449] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.740659] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1078.741711] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ae66ae0-3473-4442-8757-2947e7638e59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.748925] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1078.748925] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d57482-fa8a-29c0-7f02-68420dc2e952" [ 1078.748925] env[70013]: _type = "Task" [ 1078.748925] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.763967] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d57482-fa8a-29c0-7f02-68420dc2e952, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.764782] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9a15cea-c338-44fd-8b02-8a83ee4b81e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.771685] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1078.771685] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526021de-ddb0-ec56-4a71-42bb3e802d1d" [ 1078.771685] env[70013]: _type = "Task" [ 1078.771685] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.780226] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526021de-ddb0-ec56-4a71-42bb3e802d1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.827030] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5286222e-b1b5-4458-5650-c75fc4e1ee72, 'name': SearchDatastore_Task, 'duration_secs': 0.011399} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.827355] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.827742] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 930ababf-5f2c-4b7e-a24d-7831fbd73d66/930ababf-5f2c-4b7e-a24d-7831fbd73d66.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1078.828132] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b11ad9fd-f9d2-48c7-9f99-16328bb48b6b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.837539] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1078.837539] env[70013]: value = "task-4231357" [ 1078.837539] env[70013]: _type = "Task" [ 1078.837539] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.847601] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231357, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.867436] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231354, 'name': CloneVM_Task} progress is 95%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.888428] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.888645] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.888883] env[70013]: DEBUG nova.network.neutron [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1078.932364] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 456c4926-2ce4-4ba8-a8a6-ca50650e19f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 1078.932928] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (dc0a4c9d-b611-453e-a900-41e280c29b95): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1810}} [ 1079.282053] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526021de-ddb0-ec56-4a71-42bb3e802d1d, 'name': SearchDatastore_Task, 'duration_secs': 0.011005} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.282999] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.283301] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. {{(pid=70013) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1079.283607] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38a761c2-d497-4d6f-bf22-6d8179be0021 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.296486] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1079.296486] env[70013]: value = "task-4231358" [ 1079.296486] env[70013]: _type = "Task" [ 1079.296486] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.307170] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231358, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.349654] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231357, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.370279] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231354, 'name': CloneVM_Task, 'duration_secs': 1.839638} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.370571] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Created linked-clone VM from snapshot [ 1079.371393] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fc6f42-5292-494b-929f-8a48dca76df7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.380418] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Uploading image ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1079.410278] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1079.410278] env[70013]: value = "vm-837204" [ 1079.410278] env[70013]: _type = "VirtualMachine" [ 1079.410278] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1079.410599] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2e2b668c-325f-4d63-b8d7-4061a0212103 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.418465] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lease: (returnval){ [ 1079.418465] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526cf545-022c-f64a-3a96-2240012d220a" [ 1079.418465] env[70013]: _type = "HttpNfcLease" [ 1079.418465] env[70013]: } obtained for exporting VM: (result){ [ 1079.418465] env[70013]: value = "vm-837204" [ 1079.418465] env[70013]: _type = "VirtualMachine" [ 1079.418465] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1079.419050] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the lease: (returnval){ [ 1079.419050] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526cf545-022c-f64a-3a96-2240012d220a" [ 1079.419050] env[70013]: _type = "HttpNfcLease" [ 1079.419050] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1079.428387] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1079.428387] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526cf545-022c-f64a-3a96-2240012d220a" [ 1079.428387] env[70013]: _type = "HttpNfcLease" [ 1079.428387] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1079.435327] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance be86d96e-15be-4c39-b05d-59da0668c0d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 1079.468864] env[70013]: DEBUG nova.network.neutron [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Updated VIF entry in instance network info cache for port 4151335a-4b19-45bd-bf10-69e4960fe456. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1079.469236] env[70013]: DEBUG nova.network.neutron [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Updating instance_info_cache with network_info: [{"id": "4151335a-4b19-45bd-bf10-69e4960fe456", "address": "fa:16:3e:2c:5d:43", "network": {"id": "a6808f1c-7ccf-4a81-9bac-cdbab50a0a2a", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1787810678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940735ed506246c78bf5ede867e03751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4151335a-4b", "ovs_interfaceid": "4151335a-4b19-45bd-bf10-69e4960fe456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.610070] env[70013]: INFO nova.network.neutron [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Port 2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1079.610448] env[70013]: DEBUG nova.network.neutron [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [{"id": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "address": "fa:16:3e:46:3e:8d", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e38f8cd-84", "ovs_interfaceid": "3e38f8cd-84a0-48be-9215-3dec9c9ced94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.765368] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.765695] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.765951] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.766259] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.766971] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.768944] env[70013]: INFO nova.compute.manager [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Terminating instance [ 1079.807449] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231358, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.848849] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231357, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576274} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.849144] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 930ababf-5f2c-4b7e-a24d-7831fbd73d66/930ababf-5f2c-4b7e-a24d-7831fbd73d66.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1079.849369] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1079.849665] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2fb0bfb-5232-4048-935f-069c20318591 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.858311] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1079.858311] env[70013]: value = "task-4231360" [ 1079.858311] env[70013]: _type = "Task" [ 1079.858311] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.867733] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231360, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.927575] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1079.927575] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526cf545-022c-f64a-3a96-2240012d220a" [ 1079.927575] env[70013]: _type = "HttpNfcLease" [ 1079.927575] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1079.927950] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1079.927950] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526cf545-022c-f64a-3a96-2240012d220a" [ 1079.927950] env[70013]: _type = "HttpNfcLease" [ 1079.927950] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1079.928675] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0806d3-a54d-4416-a913-b60b83cd1059 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.936284] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a3c4fe-a11c-e036-9d73-e0288d0ff449/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1079.936460] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a3c4fe-a11c-e036-9d73-e0288d0ff449/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1079.938230] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance e3b37310-262a-4852-89b4-4d1b1d6f5776 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1794}} [ 1079.938469] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 20 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1079.938615] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4416MB phys_disk=200GB used_disk=19GB total_vcpus=48 used_vcpus=20 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1079.993600] env[70013]: DEBUG oslo_concurrency.lockutils [req-ef6525a8-4921-4dc7-b59e-fb270013b2f8 req-0c9ffd02-9d93-427b-86bf-fe2753816098 service nova] Releasing lock "refresh_cache-addf8ba9-142c-4458-8586-b06e52118d71" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.113562] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.122477] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7752d2a5-1d29-4398-8f91-7c4eb36fb73a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.276379] env[70013]: DEBUG nova.compute.manager [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1080.276379] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.278934] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22accae-71cd-42dc-82c1-705768c56705 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.293447] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.293891] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-300933f8-7828-4553-bdb4-ecd03d496ce7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.312516] env[70013]: DEBUG oslo_vmware.api [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1080.312516] env[70013]: value = "task-4231361" [ 1080.312516] env[70013]: _type = "Task" [ 1080.312516] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.312786] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231358, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.324570] env[70013]: DEBUG oslo_vmware.api [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.374661] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231360, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0672} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.375101] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1080.378966] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ad8bd7-802b-4e08-a3de-86ffa445c2dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.406902] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 930ababf-5f2c-4b7e-a24d-7831fbd73d66/930ababf-5f2c-4b7e-a24d-7831fbd73d66.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.410810] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb240532-ca6a-462a-a342-e029cd1c791f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.437472] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1080.437472] env[70013]: value = "task-4231362" [ 1080.437472] env[70013]: _type = "Task" [ 1080.437472] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.443328] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8143ebb-2019-49fb-9166-a5e2dd264aa9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.453960] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231362, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.455063] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a108433-eed7-4be4-be00-8f277a49fc8d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.489627] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7a2bc8-eb4c-4c76-bdeb-a57f66060676 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.498104] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c1fee6-40dd-486c-9083-9ca4e60db48a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.514061] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.621071] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56ada936-9821-47a0-a0ac-cebc008c20c5 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6-7b19b577-689c-4e1d-bdb0-bab33f15bd8f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.146s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1080.742558] env[70013]: DEBUG nova.compute.manager [req-b58e08f3-e434-44d2-8a52-17eb5d24f4f4 req-81629a59-dfa6-444b-9f8e-4c9785bc7f54 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-vif-deleted-2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1080.810350] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231358, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.131535} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.810715] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. [ 1080.811706] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b5d7a3-fff1-4669-a997-9ac60a094f63 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.844126] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.847806] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6662054d-4388-4a14-ada9-41bf529ec012 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.865684] env[70013]: DEBUG oslo_vmware.api [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231361, 'name': PowerOffVM_Task, 'duration_secs': 0.305445} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.866131] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.867053] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.867481] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8117acb-9ade-4be5-8c94-ebc51f3d4eb4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.874114] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1080.874114] env[70013]: value = "task-4231363" [ 1080.874114] env[70013]: _type = "Task" [ 1080.874114] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.885853] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231363, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.947712] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231362, 'name': ReconfigVM_Task, 'duration_secs': 0.342824} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.948127] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 930ababf-5f2c-4b7e-a24d-7831fbd73d66/930ababf-5f2c-4b7e-a24d-7831fbd73d66.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.948862] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae5ce9b2-4901-421f-a059-00df6eeaf029 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.957282] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1080.957282] env[70013]: value = "task-4231365" [ 1080.957282] env[70013]: _type = "Task" [ 1080.957282] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.975486] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231365, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.993997] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.994740] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.994870] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleting the datastore file [datastore1] 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.995370] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6afda2c0-6601-4f85-a668-85fc61f17e6a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.004596] env[70013]: DEBUG oslo_vmware.api [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1081.004596] env[70013]: value = "task-4231366" [ 1081.004596] env[70013]: _type = "Task" [ 1081.004596] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.016248] env[70013]: DEBUG oslo_vmware.api [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.017325] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1081.386465] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231363, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.469478] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231365, 'name': Rename_Task, 'duration_secs': 0.15812} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.471204] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1081.471204] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fdc00f15-ae1f-4393-92d0-629b1a443290 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.479294] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1081.479294] env[70013]: value = "task-4231367" [ 1081.479294] env[70013]: _type = "Task" [ 1081.479294] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.491344] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231367, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.516496] env[70013]: DEBUG oslo_vmware.api [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.481688} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.516894] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.517187] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.517415] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.517702] env[70013]: INFO nova.compute.manager [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1081.518033] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1081.518266] env[70013]: DEBUG nova.compute.manager [-] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1081.518450] env[70013]: DEBUG nova.network.neutron [-] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1081.524303] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1081.524525] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.685s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.524977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.682s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.525334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.527768] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.741s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.528320] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.530725] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.988s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.533041] env[70013]: INFO nova.compute.claims [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1081.560675] env[70013]: INFO nova.scheduler.client.report [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Deleted allocations for instance 181238eb-cb0c-4740-9896-ca745b53ebe8 [ 1081.565666] env[70013]: INFO nova.scheduler.client.report [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted allocations for instance ec7d0e9a-4a73-4681-9395-a3c963772f62 [ 1081.600240] env[70013]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 could not be found.", "detail": ""}} {{(pid=70013) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1081.600578] env[70013]: DEBUG nova.network.neutron [-] Unable to show port 2c185ee8-90c3-4a81-b3d6-0b3f21e4fa84 as it no longer exists. {{(pid=70013) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1081.886772] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231363, 'name': ReconfigVM_Task, 'duration_secs': 0.79756} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.887251] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Reconfigured VM instance instance-0000005d to attach disk [datastore2] c57e876a-8696-4ce5-b533-7580f484414e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.888364] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0421bb-cb9e-49c6-9e32-0af998b6cbf3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.917745] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a1b8843-5c94-4c49-aa8e-1e896a43afe3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.939619] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1081.939619] env[70013]: value = "task-4231368" [ 1081.939619] env[70013]: _type = "Task" [ 1081.939619] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.954775] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231368, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.994363] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231367, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.076884] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e9015792-e8c4-4ffc-a7c3-bdf79670ce37 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "ec7d0e9a-4a73-4681-9395-a3c963772f62" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.215s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.081075] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0434ce37-9589-407d-b8e3-d453d15966e1 tempest-ServersTestMultiNic-1235703585 tempest-ServersTestMultiNic-1235703585-project-member] Lock "181238eb-cb0c-4740-9896-ca745b53ebe8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.516s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.450992] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231368, 'name': ReconfigVM_Task, 'duration_secs': 0.341026} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.451908] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.452624] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a66be88-95c5-4fe5-8b46-246f29888e6f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.460915] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1082.460915] env[70013]: value = "task-4231369" [ 1082.460915] env[70013]: _type = "Task" [ 1082.460915] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.470831] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.491125] env[70013]: DEBUG oslo_vmware.api [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231367, 'name': PowerOnVM_Task, 'duration_secs': 0.542006} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.491854] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1082.493785] env[70013]: INFO nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Took 8.72 seconds to spawn the instance on the hypervisor. [ 1082.493785] env[70013]: DEBUG nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1082.493785] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d7079a-2d82-4aa7-adf2-8f92617b8594 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.630895] env[70013]: DEBUG nova.network.neutron [-] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.845724] env[70013]: DEBUG nova.compute.manager [req-4fa110f9-3c9f-440a-8a7b-a410aa36f761 req-5171bb14-bd15-4fb4-a431-ee3d3ae06c94 service nova] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Received event network-vif-deleted-3e38f8cd-84a0-48be-9215-3dec9c9ced94 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1082.944251] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34684def-fa3a-464a-8f7d-b6e38daf9431 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.955035] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59362437-fb31-4b0c-9344-fce6aa22c497 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.993613] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b0bcd7-f38c-4be8-a393-79f609bdadb0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.000190] env[70013]: DEBUG oslo_vmware.api [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231369, 'name': PowerOnVM_Task, 'duration_secs': 0.519733} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.000956] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.010524] env[70013]: DEBUG nova.compute.manager [None req-990ea71c-9e23-4510-ad7d-bea6b437c89d tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1083.012257] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716905a7-29d6-48f8-b3c6-3555bb4d55c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.022209] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cc5e03-98de-486b-ae71-b0c463a3558e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.027678] env[70013]: INFO nova.compute.manager [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Took 22.84 seconds to build instance. [ 1083.038370] env[70013]: DEBUG nova.compute.provider_tree [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.134707] env[70013]: INFO nova.compute.manager [-] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Took 1.62 seconds to deallocate network for instance. [ 1083.325667] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.326078] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.530351] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1d081048-8c61-444c-9e8c-bb73d3559f50 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.353s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.546656] env[70013]: DEBUG nova.scheduler.client.report [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1083.642665] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.829136] env[70013]: DEBUG nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1084.057025] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.057025] env[70013]: DEBUG nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1084.058620] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.916s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.058979] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.061590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.855s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.063960] env[70013]: INFO nova.compute.claims [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1084.203459] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.203893] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.204140] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.204362] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.204621] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.206788] env[70013]: INFO nova.compute.manager [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Terminating instance [ 1084.350413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.569281] env[70013]: DEBUG nova.compute.utils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1084.576313] env[70013]: DEBUG nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1084.576961] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64ec9772-0b36-46bc-b0b3-0b1a00c11c51 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 36.379s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.577818] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 14.595s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.578333] env[70013]: INFO nova.compute.manager [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Unshelving [ 1084.711036] env[70013]: DEBUG nova.compute.manager [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1084.711318] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1084.712408] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc029d07-c8c0-41d7-991c-6b0b003c63ed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.721494] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1084.721758] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae5d1e4d-0034-4770-b593-a1dc529eb012 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.730214] env[70013]: DEBUG oslo_vmware.api [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1084.730214] env[70013]: value = "task-4231370" [ 1084.730214] env[70013]: _type = "Task" [ 1084.730214] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.740319] env[70013]: DEBUG oslo_vmware.api [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231370, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.801707] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.801956] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.077559] env[70013]: DEBUG nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1085.243629] env[70013]: DEBUG oslo_vmware.api [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231370, 'name': PowerOffVM_Task, 'duration_secs': 0.326829} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.243944] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.244135] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1085.244431] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9337ef5a-834d-4de3-acaf-22ab8720f882 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.304588] env[70013]: DEBUG nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1085.319650] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1085.319930] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1085.320173] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleting the datastore file [datastore1] 930ababf-5f2c-4b7e-a24d-7831fbd73d66 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.320493] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab8fe1fb-6172-461d-af68-0471bf1f6b3f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.332270] env[70013]: DEBUG oslo_vmware.api [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for the task: (returnval){ [ 1085.332270] env[70013]: value = "task-4231372" [ 1085.332270] env[70013]: _type = "Task" [ 1085.332270] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.343030] env[70013]: DEBUG oslo_vmware.api [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231372, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.450962] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eced59fa-c91d-4d06-a255-c9fd065ef87f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.459634] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33a0c7e-7686-4352-a3b4-354aee6559fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.492509] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9ff8c0-11b0-4893-bfa0-8e1062f53c9b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.501263] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7ceb41-5c43-415b-a523-3fb8af28e8a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.516364] env[70013]: DEBUG nova.compute.provider_tree [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.591059] env[70013]: DEBUG nova.compute.utils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1085.826506] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.843549] env[70013]: DEBUG oslo_vmware.api [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Task: {'id': task-4231372, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.487282} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.843811] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.844081] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1085.844288] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1085.844471] env[70013]: INFO nova.compute.manager [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1085.844778] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1085.844988] env[70013]: DEBUG nova.compute.manager [-] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1085.845104] env[70013]: DEBUG nova.network.neutron [-] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1086.019797] env[70013]: DEBUG nova.scheduler.client.report [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1086.089921] env[70013]: DEBUG nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1086.095325] env[70013]: INFO nova.virt.block_device [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Booting with volume 5a787b14-e921-4a07-a8c6-06fd861030f1 at /dev/sdb [ 1086.122406] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1086.122685] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.122871] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1086.123656] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.123922] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1086.124118] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1086.124385] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1086.124556] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1086.124836] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1086.124995] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1086.125251] env[70013]: DEBUG nova.virt.hardware [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1086.126572] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab54c5d-fbbe-4d50-b7a2-d6af3840d10b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.138328] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d654643b-8e4e-44ef-ab65-0d5c6e631e9b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.143764] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3830dd2b-9758-4620-a14f-9292e9d923ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.151036] env[70013]: DEBUG nova.compute.manager [req-9f96b778-6e7f-493f-a0db-fc4eefb42e81 req-7b1a96f5-f667-40d2-b6ee-eee6a6c19fef service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Received event network-vif-deleted-c617a733-3882-4dd2-842d-f2be129d99ae {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1086.151036] env[70013]: INFO nova.compute.manager [req-9f96b778-6e7f-493f-a0db-fc4eefb42e81 req-7b1a96f5-f667-40d2-b6ee-eee6a6c19fef service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Neutron deleted interface c617a733-3882-4dd2-842d-f2be129d99ae; detaching it from the instance and deleting it from the info cache [ 1086.151196] env[70013]: DEBUG nova.network.neutron [req-9f96b778-6e7f-493f-a0db-fc4eefb42e81 req-7b1a96f5-f667-40d2-b6ee-eee6a6c19fef service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.161621] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.167169] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Creating folder: Project (735ed752ddd649569c5d57706d1d5efd). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1086.169559] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69341afe-c9fb-4707-8727-30e315a652f5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.174157] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529f74a6-e495-4126-b58a-e685bfa3e265 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.195169] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Created folder: Project (735ed752ddd649569c5d57706d1d5efd) in parent group-v836999. [ 1086.195402] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Creating folder: Instances. Parent ref: group-v837205. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1086.195641] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26572ed6-3170-4aec-91b9-a82fe04f7b4f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.207424] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aef239b6-e693-4de8-a9d0-983db9a4c891 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.218651] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aaa3cbd-a6cf-4bc0-aceb-6bc63e87470b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.229146] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Created folder: Instances in parent group-v837205. [ 1086.229398] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1086.229976] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1086.230850] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aebc5c88-469e-4fd5-a4df-77200572eafd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.252981] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdbc799-4b86-4a3a-8ea2-89bfca2f2af8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.257394] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.257394] env[70013]: value = "task-4231375" [ 1086.257394] env[70013]: _type = "Task" [ 1086.257394] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.262583] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf3f60c-87a7-4c2e-bf81-6fb135b04e28 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.270962] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231375, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.276897] env[70013]: DEBUG nova.virt.block_device [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating existing volume attachment record: 06422efa-d203-436d-97c1-e3d13b343c13 {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1086.525053] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.525630] env[70013]: DEBUG nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1086.528563] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.994s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.528787] env[70013]: DEBUG nova.objects.instance [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1086.634473] env[70013]: DEBUG nova.network.neutron [-] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.669148] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5784add9-7a6a-4b15-8df0-06d9e0d0bbf6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.679128] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1577691a-40d8-4f1e-9909-a9f021e736aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.717877] env[70013]: DEBUG nova.compute.manager [req-9f96b778-6e7f-493f-a0db-fc4eefb42e81 req-7b1a96f5-f667-40d2-b6ee-eee6a6c19fef service nova] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Detach interface failed, port_id=c617a733-3882-4dd2-842d-f2be129d99ae, reason: Instance 930ababf-5f2c-4b7e-a24d-7831fbd73d66 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1086.769906] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231375, 'name': CreateVM_Task, 'duration_secs': 0.403854} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.770187] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1086.770876] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.771023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.771408] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1086.771835] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04bf1042-8438-4d0e-b8c6-164aa8cf1652 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.777101] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1086.777101] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521b4e84-b20d-dfcd-a3a4-7d79007fdc75" [ 1086.777101] env[70013]: _type = "Task" [ 1086.777101] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.786611] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521b4e84-b20d-dfcd-a3a4-7d79007fdc75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.033909] env[70013]: DEBUG nova.compute.utils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1087.038225] env[70013]: DEBUG nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1087.038509] env[70013]: DEBUG nova.network.neutron [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1087.092113] env[70013]: DEBUG nova.policy [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba91f111e5c34675b39e19972f910c68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97974a99400548cf86f0f077736481b5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1087.137449] env[70013]: INFO nova.compute.manager [-] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Took 1.29 seconds to deallocate network for instance. [ 1087.288563] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521b4e84-b20d-dfcd-a3a4-7d79007fdc75, 'name': SearchDatastore_Task, 'duration_secs': 0.014683} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.288952] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.289327] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.289633] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.289806] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.290124] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.290509] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-319d56f3-f898-4f28-83bf-b2bfd0a5b990 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.300655] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.300878] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1087.301741] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c941e4a3-c385-4d3b-aa8a-86cb3ed7e2a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.308254] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1087.308254] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3d364-2224-d1fe-7bc7-4d201c0222d3" [ 1087.308254] env[70013]: _type = "Task" [ 1087.308254] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.317862] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3d364-2224-d1fe-7bc7-4d201c0222d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.420990] env[70013]: DEBUG nova.network.neutron [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Successfully created port: a1c2291b-1a1f-484e-98f5-47d6947da8a8 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1087.541789] env[70013]: DEBUG nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1087.545554] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56e64bb7-f6db-4c53-aa32-30c2674a0840 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1087.546781] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.761s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.548315] env[70013]: INFO nova.compute.claims [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.644602] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.820065] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3d364-2224-d1fe-7bc7-4d201c0222d3, 'name': SearchDatastore_Task, 'duration_secs': 0.013717} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.820918] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-900dfd29-3b13-43eb-bf25-f5f6ccfe1334 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.827345] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1087.827345] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52281331-db83-dfb3-c484-53c941637fab" [ 1087.827345] env[70013]: _type = "Task" [ 1087.827345] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.836237] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52281331-db83-dfb3-c484-53c941637fab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.338963] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52281331-db83-dfb3-c484-53c941637fab, 'name': SearchDatastore_Task, 'duration_secs': 0.014928} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.339313] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.339626] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1088.339925] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-787bb8d8-e10d-43ef-a8cf-bea7ecfcf802 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.347722] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1088.347722] env[70013]: value = "task-4231379" [ 1088.347722] env[70013]: _type = "Task" [ 1088.347722] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.357393] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231379, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.555662] env[70013]: DEBUG nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1088.583223] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1088.583459] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1088.583647] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1088.584035] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1088.584226] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1088.584383] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1088.584630] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1088.585032] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1088.585032] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1088.585257] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1088.585374] env[70013]: DEBUG nova.virt.hardware [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1088.586312] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb828b36-1cab-474e-90de-138f78e3f32b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.596407] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f152ac76-69f7-4a46-8032-2fced561e468 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.860674] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231379, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.893692] env[70013]: DEBUG nova.compute.manager [req-d7fca519-0279-4b5b-81cf-a86923613dc9 req-bc2161fa-259e-45bb-8aba-bfbc1cc81efe service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Received event network-vif-plugged-a1c2291b-1a1f-484e-98f5-47d6947da8a8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1088.895249] env[70013]: DEBUG oslo_concurrency.lockutils [req-d7fca519-0279-4b5b-81cf-a86923613dc9 req-bc2161fa-259e-45bb-8aba-bfbc1cc81efe service nova] Acquiring lock "be86d96e-15be-4c39-b05d-59da0668c0d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.895249] env[70013]: DEBUG oslo_concurrency.lockutils [req-d7fca519-0279-4b5b-81cf-a86923613dc9 req-bc2161fa-259e-45bb-8aba-bfbc1cc81efe service nova] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.895249] env[70013]: DEBUG oslo_concurrency.lockutils [req-d7fca519-0279-4b5b-81cf-a86923613dc9 req-bc2161fa-259e-45bb-8aba-bfbc1cc81efe service nova] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.895249] env[70013]: DEBUG nova.compute.manager [req-d7fca519-0279-4b5b-81cf-a86923613dc9 req-bc2161fa-259e-45bb-8aba-bfbc1cc81efe service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] No waiting events found dispatching network-vif-plugged-a1c2291b-1a1f-484e-98f5-47d6947da8a8 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1088.895691] env[70013]: WARNING nova.compute.manager [req-d7fca519-0279-4b5b-81cf-a86923613dc9 req-bc2161fa-259e-45bb-8aba-bfbc1cc81efe service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Received unexpected event network-vif-plugged-a1c2291b-1a1f-484e-98f5-47d6947da8a8 for instance with vm_state building and task_state spawning. [ 1088.976202] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f1d68d-cc61-43cf-923e-a3c69ae264fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.987731] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3847392f-7afa-4a2b-bf30-90bc6e10a669 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.028439] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be94623-5aa3-4a55-8265-e35e1380ae26 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.032377] env[70013]: DEBUG nova.network.neutron [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Successfully updated port: a1c2291b-1a1f-484e-98f5-47d6947da8a8 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1089.042298] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5569bec-51e4-4297-9ef1-8ddd9006b348 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.058421] env[70013]: DEBUG nova.compute.provider_tree [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.137154] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a3c4fe-a11c-e036-9d73-e0288d0ff449/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1089.138436] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bff3a02-26e9-4961-a09e-2142d38201a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.145317] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a3c4fe-a11c-e036-9d73-e0288d0ff449/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1089.145491] env[70013]: ERROR oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a3c4fe-a11c-e036-9d73-e0288d0ff449/disk-0.vmdk due to incomplete transfer. [ 1089.145737] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f2cd5cee-71d4-4dd9-a71b-2ebf4e86be7a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.155172] env[70013]: DEBUG oslo_vmware.rw_handles [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a3c4fe-a11c-e036-9d73-e0288d0ff449/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1089.155444] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Uploaded image ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1089.158201] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1089.158562] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5e027785-432e-4b86-a276-04e5f8d2a81a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.167295] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1089.167295] env[70013]: value = "task-4231381" [ 1089.167295] env[70013]: _type = "Task" [ 1089.167295] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.180248] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231381, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.361511] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231379, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634486} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.362131] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1089.362475] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.362829] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ef1ef7b-f7b0-4ac9-b35d-bae2e6a68a97 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.371497] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1089.371497] env[70013]: value = "task-4231382" [ 1089.371497] env[70013]: _type = "Task" [ 1089.371497] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.381179] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231382, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.536455] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "refresh_cache-be86d96e-15be-4c39-b05d-59da0668c0d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.536687] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquired lock "refresh_cache-be86d96e-15be-4c39-b05d-59da0668c0d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1089.536901] env[70013]: DEBUG nova.network.neutron [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1089.562029] env[70013]: DEBUG nova.scheduler.client.report [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1089.678733] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231381, 'name': Destroy_Task, 'duration_secs': 0.367799} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.678999] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Destroyed the VM [ 1089.679226] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1089.679491] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8fb25e3a-f60d-47b9-a4e0-ebc7bb7a0076 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.687248] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1089.687248] env[70013]: value = "task-4231383" [ 1089.687248] env[70013]: _type = "Task" [ 1089.687248] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.695740] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231383, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.883994] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231382, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071647} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.884326] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.885221] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b1c988-259f-42e9-939a-52417c09ceaf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.905020] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.905327] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5602aed5-dc00-416c-bd4d-aec9283f550d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.925144] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1089.925144] env[70013]: value = "task-4231384" [ 1089.925144] env[70013]: _type = "Task" [ 1089.925144] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.928529] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Acquiring lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.928786] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1089.928955] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Acquiring lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.929156] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1089.929325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.931550] env[70013]: INFO nova.compute.manager [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Terminating instance [ 1089.937040] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231384, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.066327] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.066878] env[70013]: DEBUG nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1090.070038] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.428s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.070288] env[70013]: DEBUG nova.objects.instance [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'resources' on Instance uuid 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.093040] env[70013]: DEBUG nova.network.neutron [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1090.198702] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231383, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.251749] env[70013]: DEBUG nova.network.neutron [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Updating instance_info_cache with network_info: [{"id": "a1c2291b-1a1f-484e-98f5-47d6947da8a8", "address": "fa:16:3e:f2:45:7a", "network": {"id": "c99489be-2cd7-4cc1-ae2f-1b8a4ac0c8f8", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1382879150-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97974a99400548cf86f0f077736481b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9722ea4d-e4a5-48fc-b759-5c4c4796b1ef", "external-id": "nsx-vlan-transportzone-924", "segmentation_id": 924, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c2291b-1a", "ovs_interfaceid": "a1c2291b-1a1f-484e-98f5-47d6947da8a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.436880] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.438848] env[70013]: DEBUG nova.compute.manager [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1090.439117] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.439414] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-185142cc-a6c7-4ac2-b390-75d7936ada7f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.448352] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1090.448352] env[70013]: value = "task-4231385" [ 1090.448352] env[70013]: _type = "Task" [ 1090.448352] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.457511] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.573986] env[70013]: DEBUG nova.compute.utils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1090.575806] env[70013]: DEBUG nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1090.576079] env[70013]: DEBUG nova.network.neutron [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1090.694596] env[70013]: DEBUG nova.policy [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfa9f992483942f689f616b60a27b00d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0896b76438d9456fb1b9fa2f39d0decc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1090.704307] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231383, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.754818] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Releasing lock "refresh_cache-be86d96e-15be-4c39-b05d-59da0668c0d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1090.755239] env[70013]: DEBUG nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Instance network_info: |[{"id": "a1c2291b-1a1f-484e-98f5-47d6947da8a8", "address": "fa:16:3e:f2:45:7a", "network": {"id": "c99489be-2cd7-4cc1-ae2f-1b8a4ac0c8f8", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1382879150-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97974a99400548cf86f0f077736481b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9722ea4d-e4a5-48fc-b759-5c4c4796b1ef", "external-id": "nsx-vlan-transportzone-924", "segmentation_id": 924, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c2291b-1a", "ovs_interfaceid": "a1c2291b-1a1f-484e-98f5-47d6947da8a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1090.755773] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:45:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9722ea4d-e4a5-48fc-b759-5c4c4796b1ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1c2291b-1a1f-484e-98f5-47d6947da8a8', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1090.763869] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Creating folder: Project (97974a99400548cf86f0f077736481b5). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1090.766958] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cb3c916-3e29-4037-b624-e152378ab684 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.781199] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Created folder: Project (97974a99400548cf86f0f077736481b5) in parent group-v836999. [ 1090.781453] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Creating folder: Instances. Parent ref: group-v837210. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1090.784707] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca4e9e03-a63d-4b23-acdd-60a14dbd4526 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.797134] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Created folder: Instances in parent group-v837210. [ 1090.797401] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1090.797616] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1090.797829] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e90f7ca0-d6fb-4dcb-bf13-51e453bfd91e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.821442] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1090.821442] env[70013]: value = "task-4231388" [ 1090.821442] env[70013]: _type = "Task" [ 1090.821442] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.831983] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231388, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.928634] env[70013]: DEBUG nova.compute.manager [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Received event network-changed-a1c2291b-1a1f-484e-98f5-47d6947da8a8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1090.928936] env[70013]: DEBUG nova.compute.manager [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Refreshing instance network info cache due to event network-changed-a1c2291b-1a1f-484e-98f5-47d6947da8a8. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1090.930972] env[70013]: DEBUG oslo_concurrency.lockutils [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] Acquiring lock "refresh_cache-be86d96e-15be-4c39-b05d-59da0668c0d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.930972] env[70013]: DEBUG oslo_concurrency.lockutils [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] Acquired lock "refresh_cache-be86d96e-15be-4c39-b05d-59da0668c0d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1090.930972] env[70013]: DEBUG nova.network.neutron [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Refreshing network info cache for port a1c2291b-1a1f-484e-98f5-47d6947da8a8 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1090.942029] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231384, 'name': ReconfigVM_Task, 'duration_secs': 0.870621} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.942412] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.943714] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6602b6b7-1160-4cea-8fa7-98b4f00b01a2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.958069] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1090.958069] env[70013]: value = "task-4231389" [ 1090.958069] env[70013]: _type = "Task" [ 1090.958069] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.966556] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231385, 'name': PowerOffVM_Task, 'duration_secs': 0.313327} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.969629] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1090.969884] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1090.970118] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837178', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'name': 'volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'serial': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1090.971169] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e939c87-9b68-44ba-a0a2-da19353a3691 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.978335] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231389, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.000954] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18c6a92-c2be-4c61-873b-457ac6982b49 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.008823] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbf0961-1b09-4993-acf6-026379e81505 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.030831] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ecf01bc-8343-464d-8943-bdcd89fa3f07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.048411] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] The volume has not been displaced from its original location: [datastore1] volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e/volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1091.053792] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Reconfiguring VM instance instance-00000059 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1091.056962] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0342eabf-0b8c-45fe-8007-94d535c0e2d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.079717] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1091.079717] env[70013]: value = "task-4231390" [ 1091.079717] env[70013]: _type = "Task" [ 1091.079717] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.087380] env[70013]: DEBUG nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1091.099049] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.120195] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd9d4ab-e3c7-462b-8ec5-e367022a13bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.130563] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a31c84-7aea-4ed8-b3e3-a3718d04a589 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.165787] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1a8072-9bf7-4546-b67c-dcec8bb4270f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.174500] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d16f16-72d6-492d-ba7c-768087ffddbe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.179920] env[70013]: DEBUG nova.network.neutron [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Successfully created port: f1b01fa4-dc57-40af-958b-ba24a089ab8d {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1091.194204] env[70013]: DEBUG nova.compute.provider_tree [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.204854] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231383, 'name': RemoveSnapshot_Task, 'duration_secs': 1.184001} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.205780] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1091.206088] env[70013]: DEBUG nova.compute.manager [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1091.207183] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a87377-917f-4120-984e-eb1463608df7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.334744] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231388, 'name': CreateVM_Task, 'duration_secs': 0.335401} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.334935] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1091.335669] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.335844] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.336223] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1091.336489] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26da4c60-59f2-492b-b234-103f4ece4351 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.342230] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1091.342230] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]524b5a4e-32c0-81fe-57a1-3388c17b4809" [ 1091.342230] env[70013]: _type = "Task" [ 1091.342230] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.350520] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524b5a4e-32c0-81fe-57a1-3388c17b4809, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.468620] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231389, 'name': Rename_Task, 'duration_secs': 0.174498} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.468620] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1091.468800] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f82e664b-9b05-47d2-bd7f-fe05e15e345a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.476725] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1091.476725] env[70013]: value = "task-4231391" [ 1091.476725] env[70013]: _type = "Task" [ 1091.476725] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.488319] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231391, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.598527] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231390, 'name': ReconfigVM_Task, 'duration_secs': 0.18794} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.602023] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Reconfigured VM instance instance-00000059 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1091.604115] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb5957ba-5f33-48f7-bd50-afabc5d23c70 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.621920] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1091.621920] env[70013]: value = "task-4231392" [ 1091.621920] env[70013]: _type = "Task" [ 1091.621920] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.636236] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231392, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.700516] env[70013]: DEBUG nova.scheduler.client.report [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1091.722602] env[70013]: DEBUG nova.network.neutron [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Updated VIF entry in instance network info cache for port a1c2291b-1a1f-484e-98f5-47d6947da8a8. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1091.723013] env[70013]: DEBUG nova.network.neutron [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Updating instance_info_cache with network_info: [{"id": "a1c2291b-1a1f-484e-98f5-47d6947da8a8", "address": "fa:16:3e:f2:45:7a", "network": {"id": "c99489be-2cd7-4cc1-ae2f-1b8a4ac0c8f8", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1382879150-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97974a99400548cf86f0f077736481b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9722ea4d-e4a5-48fc-b759-5c4c4796b1ef", "external-id": "nsx-vlan-transportzone-924", "segmentation_id": 924, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c2291b-1a", "ovs_interfaceid": "a1c2291b-1a1f-484e-98f5-47d6947da8a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.725088] env[70013]: INFO nova.compute.manager [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Shelve offloading [ 1091.854266] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524b5a4e-32c0-81fe-57a1-3388c17b4809, 'name': SearchDatastore_Task, 'duration_secs': 0.042647} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.854600] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1091.854860] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1091.855130] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.855278] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.855587] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1091.855750] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea6aafa4-8fa2-4414-a1c7-e2653d76489d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.867804] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1091.867998] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1091.869064] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0617c6b-6876-45b6-b437-95583aaeafda {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.876080] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1091.876080] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52541d94-ebe6-7f3d-4533-985e1fa4e163" [ 1091.876080] env[70013]: _type = "Task" [ 1091.876080] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.886145] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52541d94-ebe6-7f3d-4533-985e1fa4e163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.986980] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231391, 'name': PowerOnVM_Task} progress is 87%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.101315] env[70013]: DEBUG nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1092.132539] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1092.132815] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.132960] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1092.133175] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.133329] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1092.133478] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1092.133689] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1092.133890] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1092.134093] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1092.134269] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1092.134448] env[70013]: DEBUG nova.virt.hardware [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1092.135475] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68be46e5-5cbe-47aa-b700-e9d4a2aca749 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.144484] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231392, 'name': ReconfigVM_Task, 'duration_secs': 0.283597} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.146929] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837178', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'name': 'volume-e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e', 'serial': 'e7be3d36-ff28-48b2-936d-b40d8e49c89e'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1092.147270] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1092.148135] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fe27d6-a396-453e-a29c-59dcc11c1259 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.151857] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b56150d-8b2a-4053-be6d-ce9c7470add4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.169306] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1092.169599] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72c9793b-8205-4a9c-8804-3f558c5e66bd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.208746] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.138s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.211385] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.861s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.213213] env[70013]: INFO nova.compute.claims [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.228206] env[70013]: DEBUG oslo_concurrency.lockutils [req-4172db2b-0bcd-4c2c-893e-93799f08925b req-f72e7181-80a8-462f-b8f2-e591941823d3 service nova] Releasing lock "refresh_cache-be86d96e-15be-4c39-b05d-59da0668c0d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.231914] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.232481] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af5b10f6-a0ff-47cc-9944-026de56f211f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.238155] env[70013]: INFO nova.scheduler.client.report [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted allocations for instance 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6 [ 1092.241887] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1092.242616] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1092.242616] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Deleting the datastore file [datastore1] 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1092.247280] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83ee189d-8a97-48fe-a0c0-0faadeff9d28 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.250353] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1092.250353] env[70013]: value = "task-4231394" [ 1092.250353] env[70013]: _type = "Task" [ 1092.250353] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.256888] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for the task: (returnval){ [ 1092.256888] env[70013]: value = "task-4231395" [ 1092.256888] env[70013]: _type = "Task" [ 1092.256888] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.266081] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1092.266333] env[70013]: DEBUG nova.compute.manager [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1092.267332] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab130198-b38d-4f6c-b42c-f1c09e69bc53 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.277089] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231395, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.282439] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.282685] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1092.282949] env[70013]: DEBUG nova.network.neutron [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1092.386648] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52541d94-ebe6-7f3d-4533-985e1fa4e163, 'name': SearchDatastore_Task, 'duration_secs': 0.027773} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.388028] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.388862] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76e7f453-4272-4267-be6f-b77682f37372 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.395199] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1092.395199] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf2eed-3040-38e4-384a-e67eb65e60b4" [ 1092.395199] env[70013]: _type = "Task" [ 1092.395199] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.404992] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf2eed-3040-38e4-384a-e67eb65e60b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.488111] env[70013]: DEBUG oslo_vmware.api [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231391, 'name': PowerOnVM_Task, 'duration_secs': 0.901295} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.488388] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1092.488611] env[70013]: INFO nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Took 6.40 seconds to spawn the instance on the hypervisor. [ 1092.488815] env[70013]: DEBUG nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1092.489669] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc6076e5-854e-4f2d-9a2f-8165152a781e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.756652] env[70013]: DEBUG oslo_concurrency.lockutils [None req-c37e5dff-9cba-41b4-a65e-26c2519d4b92 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.991s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.770581] env[70013]: DEBUG oslo_vmware.api [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Task: {'id': task-4231395, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086077} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.770895] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.771095] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.771317] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.771522] env[70013]: INFO nova.compute.manager [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Took 2.33 seconds to destroy the instance on the hypervisor. [ 1092.771940] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1092.772129] env[70013]: DEBUG nova.compute.manager [-] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1092.772249] env[70013]: DEBUG nova.network.neutron [-] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1092.886413] env[70013]: DEBUG nova.objects.instance [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lazy-loading 'flavor' on Instance uuid 8810bebc-448c-4e9b-9fbb-ce9e24611dc9 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.910826] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bf2eed-3040-38e4-384a-e67eb65e60b4, 'name': SearchDatastore_Task, 'duration_secs': 0.010739} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.911203] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.911529] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] be86d96e-15be-4c39-b05d-59da0668c0d6/be86d96e-15be-4c39-b05d-59da0668c0d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1092.911857] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a627ba3-33ed-4f70-8ed9-80fba268c7e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.920952] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1092.920952] env[70013]: value = "task-4231396" [ 1092.920952] env[70013]: _type = "Task" [ 1092.920952] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.926607] env[70013]: DEBUG nova.network.neutron [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Successfully updated port: f1b01fa4-dc57-40af-958b-ba24a089ab8d {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.937092] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231396, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.009908] env[70013]: INFO nova.compute.manager [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Took 27.49 seconds to build instance. [ 1093.066383] env[70013]: DEBUG nova.compute.manager [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Received event network-vif-plugged-f1b01fa4-dc57-40af-958b-ba24a089ab8d {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1093.066850] env[70013]: DEBUG oslo_concurrency.lockutils [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] Acquiring lock "e3b37310-262a-4852-89b4-4d1b1d6f5776-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.071605] env[70013]: DEBUG oslo_concurrency.lockutils [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.071605] env[70013]: DEBUG oslo_concurrency.lockutils [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.071605] env[70013]: DEBUG nova.compute.manager [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] No waiting events found dispatching network-vif-plugged-f1b01fa4-dc57-40af-958b-ba24a089ab8d {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1093.071605] env[70013]: WARNING nova.compute.manager [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Received unexpected event network-vif-plugged-f1b01fa4-dc57-40af-958b-ba24a089ab8d for instance with vm_state building and task_state spawning. [ 1093.071605] env[70013]: DEBUG nova.compute.manager [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Received event network-changed-f1b01fa4-dc57-40af-958b-ba24a089ab8d {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1093.071605] env[70013]: DEBUG nova.compute.manager [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Refreshing instance network info cache due to event network-changed-f1b01fa4-dc57-40af-958b-ba24a089ab8d. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1093.071605] env[70013]: DEBUG oslo_concurrency.lockutils [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] Acquiring lock "refresh_cache-e3b37310-262a-4852-89b4-4d1b1d6f5776" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.071605] env[70013]: DEBUG oslo_concurrency.lockutils [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] Acquired lock "refresh_cache-e3b37310-262a-4852-89b4-4d1b1d6f5776" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.071605] env[70013]: DEBUG nova.network.neutron [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Refreshing network info cache for port f1b01fa4-dc57-40af-958b-ba24a089ab8d {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1093.389329] env[70013]: DEBUG nova.network.neutron [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.395385] env[70013]: DEBUG oslo_concurrency.lockutils [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.399027] env[70013]: DEBUG oslo_concurrency.lockutils [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquired lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.435364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-e3b37310-262a-4852-89b4-4d1b1d6f5776" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.435364] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231396, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.516930] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f6bb455c-f422-4502-a174-005bda64e7e0 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.006s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.629085] env[70013]: DEBUG nova.network.neutron [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1093.694531] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee53428d-2183-480f-9e8c-c0ded837c0b0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.704543] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3ecfd9-6b9c-4cd7-9ab5-ac04e85b4224 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.739357] env[70013]: DEBUG nova.network.neutron [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.744023] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4bbae0-b52b-44a4-a9c0-f398277f4e39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.751066] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e6b25f-5b41-40f7-8c2a-f0977a2bedc7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.768591] env[70013]: DEBUG nova.compute.provider_tree [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.900691] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1093.935376] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231396, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58122} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.935678] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] be86d96e-15be-4c39-b05d-59da0668c0d6/be86d96e-15be-4c39-b05d-59da0668c0d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1093.935906] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1093.942036] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a52665e-6184-48bc-978e-fdb0eee531b2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.949410] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1093.949410] env[70013]: value = "task-4231397" [ 1093.949410] env[70013]: _type = "Task" [ 1093.949410] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.961815] env[70013]: INFO nova.compute.manager [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Rebuilding instance [ 1093.970857] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231397, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.029111] env[70013]: DEBUG nova.network.neutron [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.035303] env[70013]: DEBUG nova.compute.manager [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1094.037408] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562b7ea0-c83e-4332-8538-7d6e1d19e6a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.146783] env[70013]: DEBUG nova.compute.manager [req-7b437f3b-be3a-4777-b746-8a1d856c5590 req-87027790-974f-4fe9-9eb4-f9042091d694 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Received event network-vif-deleted-e56ae4d8-5c16-4511-84c0-9c50ce3fa345 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1094.146783] env[70013]: INFO nova.compute.manager [req-7b437f3b-be3a-4777-b746-8a1d856c5590 req-87027790-974f-4fe9-9eb4-f9042091d694 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Neutron deleted interface e56ae4d8-5c16-4511-84c0-9c50ce3fa345; detaching it from the instance and deleting it from the info cache [ 1094.146783] env[70013]: DEBUG nova.network.neutron [req-7b437f3b-be3a-4777-b746-8a1d856c5590 req-87027790-974f-4fe9-9eb4-f9042091d694 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.245834] env[70013]: DEBUG oslo_concurrency.lockutils [req-7e61db54-440a-4111-87c3-b08ec7e693fa req-a66d001a-0163-4179-8d8f-ac462639f9a7 service nova] Releasing lock "refresh_cache-e3b37310-262a-4852-89b4-4d1b1d6f5776" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.246324] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-e3b37310-262a-4852-89b4-4d1b1d6f5776" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.246522] env[70013]: DEBUG nova.network.neutron [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.273139] env[70013]: DEBUG nova.scheduler.client.report [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1094.409016] env[70013]: DEBUG nova.network.neutron [-] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.459990] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231397, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.20686} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.463237] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1094.463553] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1094.464349] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac78330-4caa-4438-a625-ebd4d0c67d96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.467452] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4369b2-7ddb-4913-a373-a2fe690fe480 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.490671] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] be86d96e-15be-4c39-b05d-59da0668c0d6/be86d96e-15be-4c39-b05d-59da0668c0d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1094.494041] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af654a0b-40ab-42de-84ce-fb0029859547 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.509860] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1094.509860] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d723c228-0f34-4bf6-86d1-5588b6be220f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.517549] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1094.517549] env[70013]: value = "task-4231399" [ 1094.517549] env[70013]: _type = "Task" [ 1094.517549] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.530935] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231399, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.590071] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1094.590145] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1094.590597] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleting the datastore file [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.590597] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ff92ebe-972b-4b43-a45e-f72b1b7820bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.597844] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1094.597844] env[70013]: value = "task-4231400" [ 1094.597844] env[70013]: _type = "Task" [ 1094.597844] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.607800] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.651912] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e54e2b56-20fc-4530-b9ff-6b793a5371ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.663265] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff83304-1b76-495a-b741-3f7d6b3107f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.701965] env[70013]: DEBUG nova.compute.manager [req-7b437f3b-be3a-4777-b746-8a1d856c5590 req-87027790-974f-4fe9-9eb4-f9042091d694 service nova] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Detach interface failed, port_id=e56ae4d8-5c16-4511-84c0-9c50ce3fa345, reason: Instance 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1094.781130] env[70013]: DEBUG nova.network.neutron [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1094.787662] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.788322] env[70013]: DEBUG nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1094.791148] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.965s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.793030] env[70013]: INFO nova.compute.claims [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1094.907393] env[70013]: DEBUG nova.network.neutron [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.914752] env[70013]: INFO nova.compute.manager [-] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Took 2.14 seconds to deallocate network for instance. [ 1094.937318] env[70013]: DEBUG nova.network.neutron [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Updating instance_info_cache with network_info: [{"id": "f1b01fa4-dc57-40af-958b-ba24a089ab8d", "address": "fa:16:3e:76:ac:55", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1b01fa4-dc", "ovs_interfaceid": "f1b01fa4-dc57-40af-958b-ba24a089ab8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.028169] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.053042] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.053380] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c988af45-9b7b-4d1d-bd9f-11b2204ed8ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.061465] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1095.061465] env[70013]: value = "task-4231401" [ 1095.061465] env[70013]: _type = "Task" [ 1095.061465] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.072215] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.108245] env[70013]: DEBUG oslo_vmware.api [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.334604} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.108527] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.108721] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1095.108902] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1095.113591] env[70013]: DEBUG nova.compute.manager [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received event network-vif-unplugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1095.113961] env[70013]: DEBUG oslo_concurrency.lockutils [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.114218] env[70013]: DEBUG oslo_concurrency.lockutils [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.114412] env[70013]: DEBUG oslo_concurrency.lockutils [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1095.114611] env[70013]: DEBUG nova.compute.manager [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] No waiting events found dispatching network-vif-unplugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1095.114811] env[70013]: WARNING nova.compute.manager [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received unexpected event network-vif-unplugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 for instance with vm_state shelved and task_state shelving_offloading. [ 1095.115064] env[70013]: DEBUG nova.compute.manager [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received event network-changed-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1095.115204] env[70013]: DEBUG nova.compute.manager [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Refreshing instance network info cache due to event network-changed-d8aacd8d-2fc2-477b-9d82-e101e20b0132. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1095.115401] env[70013]: DEBUG oslo_concurrency.lockutils [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] Acquiring lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.115597] env[70013]: DEBUG oslo_concurrency.lockutils [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] Acquired lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.115774] env[70013]: DEBUG nova.network.neutron [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Refreshing network info cache for port d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1095.135294] env[70013]: INFO nova.scheduler.client.report [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleted allocations for instance 0ea84d0d-905c-428d-8abb-2781c817f08f [ 1095.303497] env[70013]: DEBUG nova.compute.utils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1095.305968] env[70013]: DEBUG nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1095.305968] env[70013]: DEBUG nova.network.neutron [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1095.349088] env[70013]: DEBUG nova.policy [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3bcb0d0bd1e44eb960d676070189f27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3d42e73dc0a4a06a2022d8b54e13e12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1095.410182] env[70013]: DEBUG oslo_concurrency.lockutils [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Releasing lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.410439] env[70013]: DEBUG nova.compute.manager [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Inject network info {{(pid=70013) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 1095.410703] env[70013]: DEBUG nova.compute.manager [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] network_info to inject: |[{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 1095.416268] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Reconfiguring VM instance to set the machine id {{(pid=70013) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 1095.416604] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58a6d50c-917a-40db-8bfe-5176bd79f2c0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.434436] env[70013]: DEBUG oslo_vmware.api [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1095.434436] env[70013]: value = "task-4231402" [ 1095.434436] env[70013]: _type = "Task" [ 1095.434436] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.438417] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-e3b37310-262a-4852-89b4-4d1b1d6f5776" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.438766] env[70013]: DEBUG nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Instance network_info: |[{"id": "f1b01fa4-dc57-40af-958b-ba24a089ab8d", "address": "fa:16:3e:76:ac:55", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1b01fa4-dc", "ovs_interfaceid": "f1b01fa4-dc57-40af-958b-ba24a089ab8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1095.443026] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:ac:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac7039c0-3374-4c08-87fc-af2449b48b02', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1b01fa4-dc57-40af-958b-ba24a089ab8d', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1095.446597] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1095.447304] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1095.447943] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11e8ce57-c4df-4e92-8357-425ab076892d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.468752] env[70013]: DEBUG oslo_vmware.api [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231402, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.471519] env[70013]: INFO nova.compute.manager [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Took 0.56 seconds to detach 1 volumes for instance. [ 1095.474434] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1095.474434] env[70013]: value = "task-4231403" [ 1095.474434] env[70013]: _type = "Task" [ 1095.474434] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.474982] env[70013]: DEBUG nova.compute.manager [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Deleting volume: e7be3d36-ff28-48b2-936d-b40d8e49c89e {{(pid=70013) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 1095.486269] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231403, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.494310] env[70013]: DEBUG nova.objects.instance [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lazy-loading 'flavor' on Instance uuid 8810bebc-448c-4e9b-9fbb-ce9e24611dc9 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.533629] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231399, 'name': ReconfigVM_Task, 'duration_secs': 0.617413} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.534080] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Reconfigured VM instance instance-00000061 to attach disk [datastore1] be86d96e-15be-4c39-b05d-59da0668c0d6/be86d96e-15be-4c39-b05d-59da0668c0d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.534827] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9300ad1c-637e-4dfc-bde1-c7e317f4b40c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.544550] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1095.544550] env[70013]: value = "task-4231404" [ 1095.544550] env[70013]: _type = "Task" [ 1095.544550] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.561901] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231404, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.578362] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231401, 'name': PowerOffVM_Task, 'duration_secs': 0.161865} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.578748] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1095.578853] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.580768] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ca37d5-2f2f-4481-a7d7-421617800aaa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.591610] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1095.591610] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be468fb6-6918-424e-8930-8fcd49fb8109 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.637794] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1095.638119] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1095.638333] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Deleting the datastore file [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.638689] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56abfa03-612c-4831-9a4a-0b8f173bb3dc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.642287] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.648491] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1095.648491] env[70013]: value = "task-4231407" [ 1095.648491] env[70013]: _type = "Task" [ 1095.648491] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.662543] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.698750] env[70013]: DEBUG nova.network.neutron [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Successfully created port: ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1095.777607] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.777987] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.809616] env[70013]: DEBUG nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1095.928128] env[70013]: DEBUG nova.network.neutron [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updated VIF entry in instance network info cache for port d8aacd8d-2fc2-477b-9d82-e101e20b0132. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1095.928575] env[70013]: DEBUG nova.network.neutron [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.949426] env[70013]: DEBUG oslo_vmware.api [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231402, 'name': ReconfigVM_Task, 'duration_secs': 0.226211} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.949635] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-576fef60-56d0-4b25-a0b4-0d386d0c9a84 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Reconfigured VM instance to set the machine id {{(pid=70013) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 1095.987703] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231403, 'name': CreateVM_Task, 'duration_secs': 0.391552} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.991042] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1095.991956] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.992181] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.992460] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1095.992989] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a68e1ae6-d6db-4652-83d2-77e9703919ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.998489] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.998652] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquired lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.001868] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1096.001868] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]524cab26-eb2f-2f80-99c6-3c0a09cf846d" [ 1096.001868] env[70013]: _type = "Task" [ 1096.001868] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.015335] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524cab26-eb2f-2f80-99c6-3c0a09cf846d, 'name': SearchDatastore_Task, 'duration_secs': 0.009858} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.015673] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.015951] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1096.016265] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.016453] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.016685] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1096.019318] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c9240c4-7425-46d9-9df8-6742d804a538 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.028346] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1096.028545] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1096.029292] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12b4674c-3bdd-4620-8988-c32c9bb0cea2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.034768] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.038876] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1096.038876] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52381c4d-339c-abb4-d6b3-478459b03da9" [ 1096.038876] env[70013]: _type = "Task" [ 1096.038876] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.049723] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52381c4d-339c-abb4-d6b3-478459b03da9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.057804] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231404, 'name': Rename_Task, 'duration_secs': 0.193396} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.059730] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1096.060301] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41f5f139-63e8-43a5-b050-78df0d71d48d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.069786] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1096.069786] env[70013]: value = "task-4231408" [ 1096.069786] env[70013]: _type = "Task" [ 1096.069786] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.078868] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231408, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.158980] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136198} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.162173] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.162372] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.162548] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.176576] env[70013]: DEBUG nova.compute.manager [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Received event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1096.176813] env[70013]: DEBUG nova.compute.manager [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing instance network info cache due to event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1096.176964] env[70013]: DEBUG oslo_concurrency.lockutils [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] Acquiring lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.190550] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc955edd-aa82-4675-883f-36e0cd927870 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.199367] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a685fab1-34ed-4338-ab31-6aac8a6ff57c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.238619] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df675ba-3dbd-4725-84a3-4373eb37e227 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.247763] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1a601a-0ec7-4e32-96e5-7bf7f29bc2c7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.264343] env[70013]: DEBUG nova.compute.provider_tree [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.282053] env[70013]: DEBUG nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1096.431665] env[70013]: DEBUG oslo_concurrency.lockutils [req-d53c4e18-38ba-4a8c-ab92-ff4414b638bf req-ea6292d7-f729-4e1c-93c2-f8de5ef13269 service nova] Releasing lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.550797] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52381c4d-339c-abb4-d6b3-478459b03da9, 'name': SearchDatastore_Task, 'duration_secs': 0.009233} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.551674] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c210f4b4-3ba6-460c-ad4d-ac8e645bb800 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.558406] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1096.558406] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fa0e09-70e3-188c-0609-f8e82b24a461" [ 1096.558406] env[70013]: _type = "Task" [ 1096.558406] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.566883] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fa0e09-70e3-188c-0609-f8e82b24a461, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.578295] env[70013]: DEBUG oslo_vmware.api [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231408, 'name': PowerOnVM_Task, 'duration_secs': 0.477769} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.578569] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1096.578777] env[70013]: INFO nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Took 8.02 seconds to spawn the instance on the hypervisor. [ 1096.578963] env[70013]: DEBUG nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1096.579762] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f253fd-08d6-4c24-97a8-63eff5f5fcd1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.748721] env[70013]: DEBUG nova.network.neutron [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1096.768321] env[70013]: DEBUG nova.scheduler.client.report [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.806127] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.820504] env[70013]: DEBUG nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1096.847976] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1096.848295] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1096.848494] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1096.848716] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1096.848907] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1096.849104] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1096.849360] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1096.849561] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1096.849797] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1096.850018] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1096.850381] env[70013]: DEBUG nova.virt.hardware [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1096.851436] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb89ab5-da80-47b9-8721-1514f351a4b4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.863046] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09586a2a-d6fd-4248-965b-a383e9d1daf6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.071991] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52fa0e09-70e3-188c-0609-f8e82b24a461, 'name': SearchDatastore_Task, 'duration_secs': 0.013917} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.072412] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1097.072550] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] e3b37310-262a-4852-89b4-4d1b1d6f5776/e3b37310-262a-4852-89b4-4d1b1d6f5776.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1097.072861] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af71b0fa-4399-41cc-9e90-97f72cb51b6d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.079983] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1097.079983] env[70013]: value = "task-4231409" [ 1097.079983] env[70013]: _type = "Task" [ 1097.079983] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.088541] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231409, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.098642] env[70013]: INFO nova.compute.manager [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Took 21.91 seconds to build instance. [ 1097.153020] env[70013]: DEBUG nova.compute.manager [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Received event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1097.153224] env[70013]: DEBUG nova.compute.manager [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing instance network info cache due to event network-changed-38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1097.153452] env[70013]: DEBUG oslo_concurrency.lockutils [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] Acquiring lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.199020] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1097.199020] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1097.199020] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1097.199020] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1097.199020] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1097.199020] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1097.199739] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1097.200332] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1097.200692] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1097.201025] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1097.201328] env[70013]: DEBUG nova.virt.hardware [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1097.204020] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26b8bf4-d316-485d-9ef0-6130a34d32f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.213509] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e87f8a-fdb2-4988-bb7f-bb3eca1c8c2b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.233926] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1097.240383] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1097.240979] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1097.241084] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eeb14209-bd2c-413d-8027-f6b4acbd2a76 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.259724] env[70013]: DEBUG nova.network.neutron [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Successfully updated port: ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1097.271516] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1097.271516] env[70013]: value = "task-4231410" [ 1097.271516] env[70013]: _type = "Task" [ 1097.271516] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.276694] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.277305] env[70013]: DEBUG nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1097.281241] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.637s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.281505] env[70013]: DEBUG nova.objects.instance [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lazy-loading 'resources' on Instance uuid 930ababf-5f2c-4b7e-a24d-7831fbd73d66 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.289599] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231410, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.543728] env[70013]: DEBUG nova.network.neutron [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.593665] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231409, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.601516] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d2518fe-ab46-4da0-9097-1d3045e02611 tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.428s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.765501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.765501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.765501] env[70013]: DEBUG nova.network.neutron [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1097.782563] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231410, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.785059] env[70013]: DEBUG nova.compute.utils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1097.789758] env[70013]: DEBUG nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1097.789938] env[70013]: DEBUG nova.network.neutron [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1097.839815] env[70013]: DEBUG nova.policy [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed0f82552bf43268d93688820492e44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c6094eeb5ca48dfbfbb10bc71feac67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1098.047941] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Releasing lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.048592] env[70013]: DEBUG nova.compute.manager [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Inject network info {{(pid=70013) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 1098.048986] env[70013]: DEBUG nova.compute.manager [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] network_info to inject: |[{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 1098.055722] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Reconfiguring VM instance to set the machine id {{(pid=70013) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 1098.058705] env[70013]: DEBUG oslo_concurrency.lockutils [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] Acquired lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.058705] env[70013]: DEBUG nova.network.neutron [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1098.060608] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0afd1dff-4d43-43ac-b160-e8d83a93a400 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.089882] env[70013]: DEBUG oslo_vmware.api [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1098.089882] env[70013]: value = "task-4231411" [ 1098.089882] env[70013]: _type = "Task" [ 1098.089882] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.098892] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231409, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616942} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.099573] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] e3b37310-262a-4852-89b4-4d1b1d6f5776/e3b37310-262a-4852-89b4-4d1b1d6f5776.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1098.099884] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1098.100262] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fab9df97-2905-4ed0-a8b6-d5e7f425f237 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.106925] env[70013]: DEBUG oslo_vmware.api [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231411, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.114483] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1098.114483] env[70013]: value = "task-4231412" [ 1098.114483] env[70013]: _type = "Task" [ 1098.114483] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.129015] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231412, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.154479] env[70013]: DEBUG nova.network.neutron [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Successfully created port: 1d2febcb-8ea3-4126-8de7-7bbe2f7162ec {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1098.240870] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654ee540-0710-4726-a6f9-f219dff0a814 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.249298] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d7813c-1cc0-4df6-9398-dc90c280eea2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.287824] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a9c004-9215-43ae-b425-64ce10da2b24 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.290642] env[70013]: DEBUG nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1098.301234] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231410, 'name': CreateVM_Task, 'duration_secs': 0.547681} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.302508] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6f31ad-c89a-4de9-98d0-8975be604154 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.306522] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1098.307230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.307230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.307504] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1098.307961] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dde1d98a-ec80-4312-bb39-3bf1ac355658 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.319289] env[70013]: DEBUG nova.compute.provider_tree [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.322064] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1098.322064] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d61610-91a3-1097-1cfc-811bada774c4" [ 1098.322064] env[70013]: _type = "Task" [ 1098.322064] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.330969] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d61610-91a3-1097-1cfc-811bada774c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.366208] env[70013]: DEBUG nova.network.neutron [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1098.471146] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.506020] env[70013]: DEBUG nova.compute.manager [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Received event network-vif-plugged-ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1098.506020] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] Acquiring lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.506020] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.506020] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.506020] env[70013]: DEBUG nova.compute.manager [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] No waiting events found dispatching network-vif-plugged-ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1098.506020] env[70013]: WARNING nova.compute.manager [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Received unexpected event network-vif-plugged-ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 for instance with vm_state building and task_state spawning. [ 1098.506020] env[70013]: DEBUG nova.compute.manager [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Received event network-changed-ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1098.506020] env[70013]: DEBUG nova.compute.manager [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Refreshing instance network info cache due to event network-changed-ba7e21fc-95aa-4d9c-b548-695aa34fbbb7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1098.506020] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] Acquiring lock "refresh_cache-bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.582308] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.582574] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.582804] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.582997] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.583183] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.585165] env[70013]: INFO nova.compute.manager [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Terminating instance [ 1098.593529] env[70013]: DEBUG nova.network.neutron [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Updating instance_info_cache with network_info: [{"id": "ba7e21fc-95aa-4d9c-b548-695aa34fbbb7", "address": "fa:16:3e:4a:b5:28", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba7e21fc-95", "ovs_interfaceid": "ba7e21fc-95aa-4d9c-b548-695aa34fbbb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.598825] env[70013]: DEBUG oslo_vmware.api [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231411, 'name': ReconfigVM_Task, 'duration_secs': 0.151139} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.599126] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f144fffb-c7b2-4c4d-8240-a1ad8a365443 tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Reconfigured VM instance to set the machine id {{(pid=70013) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 1098.611131] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "be86d96e-15be-4c39-b05d-59da0668c0d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.611131] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.611131] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "be86d96e-15be-4c39-b05d-59da0668c0d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.611268] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.612113] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.613792] env[70013]: INFO nova.compute.manager [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Terminating instance [ 1098.629880] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231412, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071603} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.630158] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1098.630956] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87fc728-e349-4caa-85a1-af7caef5055e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.658500] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] e3b37310-262a-4852-89b4-4d1b1d6f5776/e3b37310-262a-4852-89b4-4d1b1d6f5776.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.659152] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19cf7326-b861-4444-9a12-4a1282a0e79a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.680178] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1098.680178] env[70013]: value = "task-4231413" [ 1098.680178] env[70013]: _type = "Task" [ 1098.680178] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.691987] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231413, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.810475] env[70013]: DEBUG nova.network.neutron [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updated VIF entry in instance network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1098.810840] env[70013]: DEBUG nova.network.neutron [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.828853] env[70013]: DEBUG nova.scheduler.client.report [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1098.838916] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d61610-91a3-1097-1cfc-811bada774c4, 'name': SearchDatastore_Task, 'duration_secs': 0.071033} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.839331] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.839669] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1098.840010] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.840198] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.840434] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1098.840718] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21f8ede8-2646-49fa-a6cc-abbbb5458650 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.851496] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1098.851690] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1098.852443] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80c8f8aa-ebab-4c6c-af78-ff4246c5f8b1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.857906] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1098.857906] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b298d4-d614-1a0a-ee48-1aced40d981b" [ 1098.857906] env[70013]: _type = "Task" [ 1098.857906] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.866586] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b298d4-d614-1a0a-ee48-1aced40d981b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.088926] env[70013]: DEBUG nova.compute.manager [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1099.089251] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1099.090169] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7c1197-ae8a-40a2-8094-d5719fc9db64 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.098596] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.099135] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.099462] env[70013]: DEBUG nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Instance network_info: |[{"id": "ba7e21fc-95aa-4d9c-b548-695aa34fbbb7", "address": "fa:16:3e:4a:b5:28", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba7e21fc-95", "ovs_interfaceid": "ba7e21fc-95aa-4d9c-b548-695aa34fbbb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1099.099774] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d723c88-395e-40bd-ae29-0829f5d58718 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.101572] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] Acquired lock "refresh_cache-bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.101796] env[70013]: DEBUG nova.network.neutron [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Refreshing network info cache for port ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1099.103030] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:b5:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba7e21fc-95aa-4d9c-b548-695aa34fbbb7', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1099.110705] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1099.113778] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1099.115092] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3cc41b39-eb51-4da2-b2db-d9a5361dc7e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.134093] env[70013]: DEBUG nova.compute.manager [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1099.134368] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1099.134842] env[70013]: DEBUG oslo_vmware.api [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1099.134842] env[70013]: value = "task-4231414" [ 1099.134842] env[70013]: _type = "Task" [ 1099.134842] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.135861] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ad81af-9f10-4502-9446-e70ed846f010 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.146322] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1099.146322] env[70013]: value = "task-4231415" [ 1099.146322] env[70013]: _type = "Task" [ 1099.146322] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.155399] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.155707] env[70013]: DEBUG oslo_vmware.api [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.156499] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23425ee5-f8ad-4727-afc6-817c733e8d8b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.162493] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231415, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.164571] env[70013]: DEBUG oslo_vmware.api [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1099.164571] env[70013]: value = "task-4231416" [ 1099.164571] env[70013]: _type = "Task" [ 1099.164571] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.175320] env[70013]: DEBUG oslo_vmware.api [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231416, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.194467] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231413, 'name': ReconfigVM_Task, 'duration_secs': 0.303459} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.194698] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Reconfigured VM instance instance-00000062 to attach disk [datastore1] e3b37310-262a-4852-89b4-4d1b1d6f5776/e3b37310-262a-4852-89b4-4d1b1d6f5776.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.195244] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b721f2f-6903-45de-9d35-6eb85c47a3a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.203892] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1099.203892] env[70013]: value = "task-4231417" [ 1099.203892] env[70013]: _type = "Task" [ 1099.203892] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.214800] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231417, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.301160] env[70013]: DEBUG nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1099.314470] env[70013]: DEBUG oslo_concurrency.lockutils [req-e2092e47-88ee-40aa-98ce-6460084059be req-8736982d-dff8-4764-86be-1bbd57aeba83 service nova] Releasing lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.315074] env[70013]: DEBUG oslo_concurrency.lockutils [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] Acquired lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.315435] env[70013]: DEBUG nova.network.neutron [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Refreshing network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1099.335955] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1099.335955] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1099.335955] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1099.335955] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1099.336308] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1099.336308] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1099.336485] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1099.336672] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1099.336837] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1099.337000] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1099.337202] env[70013]: DEBUG nova.virt.hardware [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1099.338110] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.342317] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32eeee1-daa0-42ce-8a5d-b70288aa4975 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.349562] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.961s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.350048] env[70013]: DEBUG nova.objects.instance [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'pci_requests' on Instance uuid 472b8218-51c3-492e-96ff-5fa99df4cc4c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.358158] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf67d2b-e998-44c7-af11-8dbead57d189 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.382884] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b298d4-d614-1a0a-ee48-1aced40d981b, 'name': SearchDatastore_Task, 'duration_secs': 0.021406} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.384469] env[70013]: INFO nova.scheduler.client.report [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Deleted allocations for instance 930ababf-5f2c-4b7e-a24d-7831fbd73d66 [ 1099.386043] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caaa2da0-f0fa-4612-a8e2-afe4988a1d1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.393519] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1099.393519] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522bb6f2-36e7-f960-0330-ad23b9cf95e9" [ 1099.393519] env[70013]: _type = "Task" [ 1099.393519] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.402504] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522bb6f2-36e7-f960-0330-ad23b9cf95e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.650301] env[70013]: DEBUG oslo_vmware.api [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231414, 'name': PowerOffVM_Task, 'duration_secs': 0.240718} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.654221] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.654519] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1099.654955] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc890e77-db29-4d93-9ddc-71cf9160e794 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.664721] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231415, 'name': CreateVM_Task, 'duration_secs': 0.406109} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.665191] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1099.668770] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.669055] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.669433] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1099.669757] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-182045c4-c9a0-459c-89ee-51341d5c62ae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.678178] env[70013]: DEBUG oslo_vmware.api [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231416, 'name': PowerOffVM_Task, 'duration_secs': 0.213647} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.679540] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.679792] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1099.680170] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1099.680170] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52add47c-8c75-d171-dc5f-6122456ba79b" [ 1099.680170] env[70013]: _type = "Task" [ 1099.680170] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.680581] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f7e02c7-9376-4c8f-853c-0d1916a2e9f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.692055] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52add47c-8c75-d171-dc5f-6122456ba79b, 'name': SearchDatastore_Task, 'duration_secs': 0.010581} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.692460] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.692780] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1099.693080] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.714262] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231417, 'name': Rename_Task, 'duration_secs': 0.164787} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.714603] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.714897] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6268763-45e8-4eb4-9c19-9dcbec65b2ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.722577] env[70013]: DEBUG nova.network.neutron [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Successfully updated port: 1d2febcb-8ea3-4126-8de7-7bbe2f7162ec {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1099.724289] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1099.724289] env[70013]: value = "task-4231420" [ 1099.724289] env[70013]: _type = "Task" [ 1099.724289] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.738594] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231420, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.740359] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1099.740641] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1099.740796] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Deleting the datastore file [datastore1] 8810bebc-448c-4e9b-9fbb-ce9e24611dc9 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.741086] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db097c12-cf8c-40f6-99c7-0d274eb1a8e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.751305] env[70013]: DEBUG oslo_vmware.api [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for the task: (returnval){ [ 1099.751305] env[70013]: value = "task-4231421" [ 1099.751305] env[70013]: _type = "Task" [ 1099.751305] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.760993] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1099.761201] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1099.761452] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Deleting the datastore file [datastore1] be86d96e-15be-4c39-b05d-59da0668c0d6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.762912] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c8333c7-49f8-4655-8b85-5bfd75ab0449 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.767615] env[70013]: DEBUG oslo_vmware.api [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231421, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.773776] env[70013]: DEBUG oslo_vmware.api [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for the task: (returnval){ [ 1099.773776] env[70013]: value = "task-4231422" [ 1099.773776] env[70013]: _type = "Task" [ 1099.773776] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.784648] env[70013]: DEBUG oslo_vmware.api [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.854163] env[70013]: DEBUG nova.objects.instance [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'numa_topology' on Instance uuid 472b8218-51c3-492e-96ff-5fa99df4cc4c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.899854] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f20e2e22-8696-4a68-a1cc-61cc1def7c64 tempest-ServerDiskConfigTestJSON-1226251677 tempest-ServerDiskConfigTestJSON-1226251677-project-member] Lock "930ababf-5f2c-4b7e-a24d-7831fbd73d66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.696s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.908847] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522bb6f2-36e7-f960-0330-ad23b9cf95e9, 'name': SearchDatastore_Task, 'duration_secs': 0.010262} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.909163] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.909481] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1099.909841] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.910069] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1099.910330] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24793e3f-9b0b-48c0-85a8-f7c2a37db284 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.912700] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-630d9762-1b21-48e6-9db2-e04d49c7a98b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.920642] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1099.920642] env[70013]: value = "task-4231423" [ 1099.920642] env[70013]: _type = "Task" [ 1099.920642] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.926196] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1099.926425] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1099.927548] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea9c4a3b-6cbf-4f11-bafc-436fefb79fc4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.933321] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231423, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.939588] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1099.939588] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]524b0f0f-cf27-d047-a8ee-69cda1e15e69" [ 1099.939588] env[70013]: _type = "Task" [ 1099.939588] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.943706] env[70013]: DEBUG nova.network.neutron [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Updated VIF entry in instance network info cache for port ba7e21fc-95aa-4d9c-b548-695aa34fbbb7. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1099.944097] env[70013]: DEBUG nova.network.neutron [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Updating instance_info_cache with network_info: [{"id": "ba7e21fc-95aa-4d9c-b548-695aa34fbbb7", "address": "fa:16:3e:4a:b5:28", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba7e21fc-95", "ovs_interfaceid": "ba7e21fc-95aa-4d9c-b548-695aa34fbbb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.954274] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]524b0f0f-cf27-d047-a8ee-69cda1e15e69, 'name': SearchDatastore_Task, 'duration_secs': 0.010055} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.955128] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61424fd7-f3b4-416d-88c9-35d6f3fee29a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.961752] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1099.961752] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523e913c-0089-43b2-9259-819317760844" [ 1099.961752] env[70013]: _type = "Task" [ 1099.961752] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.970930] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523e913c-0089-43b2-9259-819317760844, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.092474] env[70013]: DEBUG nova.network.neutron [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updated VIF entry in instance network info cache for port 38ceae0a-4242-4587-9200-5cb36bbb5814. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1100.093757] env[70013]: DEBUG nova.network.neutron [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [{"id": "38ceae0a-4242-4587-9200-5cb36bbb5814", "address": "fa:16:3e:e6:c6:46", "network": {"id": "412aa733-5532-428e-904c-fe6ec1fc2360", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2031703349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9659cfe4a7a84e47af14354ffe4005cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38ceae0a-42", "ovs_interfaceid": "38ceae0a-4242-4587-9200-5cb36bbb5814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.227575] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.227820] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.227820] env[70013]: DEBUG nova.network.neutron [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1100.240209] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231420, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.264601] env[70013]: DEBUG oslo_vmware.api [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Task: {'id': task-4231421, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1654} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.264979] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.265111] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1100.265269] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1100.265456] env[70013]: INFO nova.compute.manager [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1100.265990] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1100.266248] env[70013]: DEBUG nova.compute.manager [-] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1100.266351] env[70013]: DEBUG nova.network.neutron [-] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1100.288446] env[70013]: DEBUG oslo_vmware.api [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Task: {'id': task-4231422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164577} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.288446] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.288446] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1100.288446] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1100.288446] env[70013]: INFO nova.compute.manager [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1100.288446] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1100.288446] env[70013]: DEBUG nova.compute.manager [-] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1100.288446] env[70013]: DEBUG nova.network.neutron [-] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1100.359195] env[70013]: INFO nova.compute.claims [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1100.432546] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231423, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.447027] env[70013]: DEBUG oslo_concurrency.lockutils [req-9b595f10-2f3a-4731-8653-c6611ed9d86a req-1a7a0375-ba22-4182-90d8-dbc678313bbf service nova] Releasing lock "refresh_cache-bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.474299] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523e913c-0089-43b2-9259-819317760844, 'name': SearchDatastore_Task, 'duration_secs': 0.008885} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.474759] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.475256] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] bf138a0d-ea1f-4c93-9c7f-e0e3309597b2/bf138a0d-ea1f-4c93-9c7f-e0e3309597b2.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1100.475604] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e8cfedf-678a-4e09-9cbf-a073520fe4dc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.485252] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1100.485252] env[70013]: value = "task-4231424" [ 1100.485252] env[70013]: _type = "Task" [ 1100.485252] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.496254] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.596242] env[70013]: DEBUG oslo_concurrency.lockutils [req-422b56c5-d385-46d9-83d7-4ea258cf7b80 req-b4744d8b-edff-4181-a1f4-d5bb5308a9a5 service nova] Releasing lock "refresh_cache-8810bebc-448c-4e9b-9fbb-ce9e24611dc9" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.605701] env[70013]: DEBUG nova.compute.manager [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Received event network-vif-plugged-1d2febcb-8ea3-4126-8de7-7bbe2f7162ec {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1100.606088] env[70013]: DEBUG oslo_concurrency.lockutils [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] Acquiring lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.607129] env[70013]: DEBUG oslo_concurrency.lockutils [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1100.607129] env[70013]: DEBUG oslo_concurrency.lockutils [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.607813] env[70013]: DEBUG nova.compute.manager [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] No waiting events found dispatching network-vif-plugged-1d2febcb-8ea3-4126-8de7-7bbe2f7162ec {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1100.608116] env[70013]: WARNING nova.compute.manager [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Received unexpected event network-vif-plugged-1d2febcb-8ea3-4126-8de7-7bbe2f7162ec for instance with vm_state building and task_state spawning. [ 1100.608375] env[70013]: DEBUG nova.compute.manager [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Received event network-changed-1d2febcb-8ea3-4126-8de7-7bbe2f7162ec {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1100.608598] env[70013]: DEBUG nova.compute.manager [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Refreshing instance network info cache due to event network-changed-1d2febcb-8ea3-4126-8de7-7bbe2f7162ec. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1100.609101] env[70013]: DEBUG oslo_concurrency.lockutils [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] Acquiring lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.742244] env[70013]: DEBUG oslo_vmware.api [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231420, 'name': PowerOnVM_Task, 'duration_secs': 0.931467} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.744021] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1100.744021] env[70013]: INFO nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Took 8.64 seconds to spawn the instance on the hypervisor. [ 1100.744021] env[70013]: DEBUG nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1100.744509] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3631f1b8-8047-40b9-b840-daadfe4f69a8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.815889] env[70013]: DEBUG nova.network.neutron [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1100.943995] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231423, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602403} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.944429] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1100.944662] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1100.944934] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0936071-127e-4b82-af03-8b6809b9b91a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.957837] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1100.957837] env[70013]: value = "task-4231425" [ 1100.957837] env[70013]: _type = "Task" [ 1100.957837] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.971782] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.997420] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231424, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.130044] env[70013]: DEBUG nova.network.neutron [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Updating instance_info_cache with network_info: [{"id": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "address": "fa:16:3e:43:93:9a", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2febcb-8e", "ovs_interfaceid": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.272388] env[70013]: INFO nova.compute.manager [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Took 24.51 seconds to build instance. [ 1101.335231] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0df81c9-2c2d-4b3e-bce9-0bbe83c9d082 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.343770] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f819928-7981-4d29-8afb-0e551088fea9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.377581] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f4388d-377c-41bf-88fb-233b2e631326 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.388171] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9febb2ca-3e7f-4a08-8aab-80ba0d886d08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.404716] env[70013]: DEBUG nova.compute.provider_tree [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.465029] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100621} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.465632] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.466756] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973746fc-19b4-4503-ab53-af3189df9dc3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.472746] env[70013]: DEBUG nova.network.neutron [-] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.490478] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.491704] env[70013]: INFO nova.compute.manager [-] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Took 1.20 seconds to deallocate network for instance. [ 1101.492104] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f6e6cef-0048-4abe-b285-6b1be99fefa9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.523044] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.741001} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.523859] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] bf138a0d-ea1f-4c93-9c7f-e0e3309597b2/bf138a0d-ea1f-4c93-9c7f-e0e3309597b2.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1101.524257] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1101.524739] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1101.524739] env[70013]: value = "task-4231426" [ 1101.524739] env[70013]: _type = "Task" [ 1101.524739] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.525061] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f08fa646-0dd6-4fba-a1ff-48e1fc9e7858 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.541026] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231426, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.541026] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1101.541026] env[70013]: value = "task-4231427" [ 1101.541026] env[70013]: _type = "Task" [ 1101.541026] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.552920] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231427, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.632948] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.633376] env[70013]: DEBUG nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Instance network_info: |[{"id": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "address": "fa:16:3e:43:93:9a", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2febcb-8e", "ovs_interfaceid": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1101.633742] env[70013]: DEBUG oslo_concurrency.lockutils [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] Acquired lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.633981] env[70013]: DEBUG nova.network.neutron [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Refreshing network info cache for port 1d2febcb-8ea3-4126-8de7-7bbe2f7162ec {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1101.635819] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:93:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d2febcb-8ea3-4126-8de7-7bbe2f7162ec', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1101.644544] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1101.645385] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1101.646326] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8bc5919d-42d8-49dc-917d-c37dacd5c210 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.670445] env[70013]: DEBUG nova.network.neutron [-] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.673377] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1101.673377] env[70013]: value = "task-4231428" [ 1101.673377] env[70013]: _type = "Task" [ 1101.673377] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.684580] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231428, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.778714] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e086aa1f-608f-4171-ad4e-f58a03e695ab tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.021s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.909599] env[70013]: DEBUG nova.scheduler.client.report [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1102.027184] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.044025] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231426, 'name': ReconfigVM_Task, 'duration_secs': 0.43116} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.044025] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6/456c4926-2ce4-4ba8-a8a6-ca50650e19f6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.046093] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d93d29f3-ef9b-46bd-ace4-6a1d7e8217fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.054357] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085577} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.056183] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1102.058159] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1102.058159] env[70013]: value = "task-4231429" [ 1102.058159] env[70013]: _type = "Task" [ 1102.058159] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.058159] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd90ca1-5d32-4946-9dd0-26a0051867f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.091023] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] bf138a0d-ea1f-4c93-9c7f-e0e3309597b2/bf138a0d-ea1f-4c93-9c7f-e0e3309597b2.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1102.094973] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-560423a9-b9f5-4641-b2a8-b86f7ff5ef99 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.111426] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231429, 'name': Rename_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.117788] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1102.117788] env[70013]: value = "task-4231430" [ 1102.117788] env[70013]: _type = "Task" [ 1102.117788] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.127436] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231430, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.174567] env[70013]: INFO nova.compute.manager [-] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Took 1.91 seconds to deallocate network for instance. [ 1102.187578] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231428, 'name': CreateVM_Task, 'duration_secs': 0.508913} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.191581] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1102.193047] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.197199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.197607] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1102.198984] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fecf563-2626-4695-b0c8-7811ac91e99e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.206371] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1102.206371] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5249b5ad-a4e0-5a16-7538-072886f3d742" [ 1102.206371] env[70013]: _type = "Task" [ 1102.206371] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.216675] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5249b5ad-a4e0-5a16-7538-072886f3d742, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.415298] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.066s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.417664] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.776s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.418237] env[70013]: DEBUG nova.objects.instance [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lazy-loading 'resources' on Instance uuid 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.437048] env[70013]: DEBUG nova.network.neutron [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Updated VIF entry in instance network info cache for port 1d2febcb-8ea3-4126-8de7-7bbe2f7162ec. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1102.437509] env[70013]: DEBUG nova.network.neutron [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Updating instance_info_cache with network_info: [{"id": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "address": "fa:16:3e:43:93:9a", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2febcb-8e", "ovs_interfaceid": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.502351] env[70013]: INFO nova.network.neutron [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating port 3fc25179-89cf-42de-b6b6-f31806bcbce3 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1102.572715] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231429, 'name': Rename_Task, 'duration_secs': 0.199208} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.576899] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1102.578827] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eef32fc8-b30d-4d2b-943a-316432128626 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.585925] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1102.585925] env[70013]: value = "task-4231431" [ 1102.585925] env[70013]: _type = "Task" [ 1102.585925] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.595878] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231431, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.636446] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231430, 'name': ReconfigVM_Task, 'duration_secs': 0.338453} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.636736] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Reconfigured VM instance instance-00000063 to attach disk [datastore1] bf138a0d-ea1f-4c93-9c7f-e0e3309597b2/bf138a0d-ea1f-4c93-9c7f-e0e3309597b2.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.638345] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8ac19d3-d5dd-4dc1-9917-7b36474e1fd1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.645876] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1102.645876] env[70013]: value = "task-4231432" [ 1102.645876] env[70013]: _type = "Task" [ 1102.645876] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.662976] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231432, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.694767] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.719217] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5249b5ad-a4e0-5a16-7538-072886f3d742, 'name': SearchDatastore_Task, 'duration_secs': 0.015176} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.719217] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.719217] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1102.719217] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.719217] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.719217] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1102.719692] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a784545-bf97-4572-9f4b-615f1f0e29e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.730869] env[70013]: DEBUG nova.compute.manager [req-d50b529e-d29a-4b36-9891-edab683f0ad5 req-bcc78634-ac9b-401e-a122-91a6fa3582da service nova] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Received event network-vif-deleted-a1c2291b-1a1f-484e-98f5-47d6947da8a8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1102.731259] env[70013]: DEBUG nova.compute.manager [req-d50b529e-d29a-4b36-9891-edab683f0ad5 req-bcc78634-ac9b-401e-a122-91a6fa3582da service nova] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Received event network-vif-deleted-38ceae0a-4242-4587-9200-5cb36bbb5814 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1102.733500] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1102.733767] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1102.734806] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-063f3834-84a4-47c0-b944-cf2d068ca267 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.741899] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "e3b37310-262a-4852-89b4-4d1b1d6f5776" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.742089] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.742339] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "e3b37310-262a-4852-89b4-4d1b1d6f5776-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.742545] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.742949] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.747228] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1102.747228] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e0bb30-579f-043d-308d-13ae54537df6" [ 1102.747228] env[70013]: _type = "Task" [ 1102.747228] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.747550] env[70013]: INFO nova.compute.manager [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Terminating instance [ 1102.762476] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e0bb30-579f-043d-308d-13ae54537df6, 'name': SearchDatastore_Task, 'duration_secs': 0.012067} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.764467] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b67bb1a-8a81-4d8f-863d-0eb10c4f3c98 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.773551] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1102.773551] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52abae6e-7747-d042-68a1-68033e403a12" [ 1102.773551] env[70013]: _type = "Task" [ 1102.773551] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.784460] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52abae6e-7747-d042-68a1-68033e403a12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.920564] env[70013]: DEBUG nova.objects.instance [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lazy-loading 'numa_topology' on Instance uuid 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.940098] env[70013]: DEBUG oslo_concurrency.lockutils [req-972795d3-b4af-4950-a9c8-5518fd1b8e26 req-b508740f-821a-4142-afe6-89354c723683 service nova] Releasing lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.097713] env[70013]: DEBUG oslo_vmware.api [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231431, 'name': PowerOnVM_Task, 'duration_secs': 0.468698} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.098087] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1103.098325] env[70013]: DEBUG nova.compute.manager [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1103.099198] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e24221-d11a-4387-96b7-065d6f56586b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.157734] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231432, 'name': Rename_Task, 'duration_secs': 0.158284} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.157912] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1103.158237] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1907033-acd0-4879-8a95-3f0f1b061217 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.166061] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1103.166061] env[70013]: value = "task-4231433" [ 1103.166061] env[70013]: _type = "Task" [ 1103.166061] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.175365] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.255864] env[70013]: DEBUG nova.compute.manager [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1103.255864] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1103.257280] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb7bf6f-2b8f-4950-b8ef-cbb64e7fcfd5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.266526] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.266732] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41629983-c265-4fd4-a512-cc8acfac045b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.275325] env[70013]: DEBUG oslo_vmware.api [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1103.275325] env[70013]: value = "task-4231434" [ 1103.275325] env[70013]: _type = "Task" [ 1103.275325] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.290941] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52abae6e-7747-d042-68a1-68033e403a12, 'name': SearchDatastore_Task, 'duration_secs': 0.014011} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.294176] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.294496] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/8dd3e9c7-3a92-49e6-883c-cda647730f5e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1103.294798] env[70013]: DEBUG oslo_vmware.api [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.295075] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da2dc0e4-e835-41ac-925d-653be1e0bd0f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.304036] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1103.304036] env[70013]: value = "task-4231435" [ 1103.304036] env[70013]: _type = "Task" [ 1103.304036] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.312461] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.423442] env[70013]: DEBUG nova.objects.base [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Object Instance<0ea84d0d-905c-428d-8abb-2781c817f08f> lazy-loaded attributes: resources,numa_topology {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1103.621858] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1103.683887] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231433, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.796663] env[70013]: DEBUG oslo_vmware.api [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231434, 'name': PowerOffVM_Task, 'duration_secs': 0.202825} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.797040] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.797821] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1103.797821] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9eb363a4-402f-4421-b48b-b097318233e1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.819393] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231435, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.881495] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1103.885157] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1103.885157] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleting the datastore file [datastore1] e3b37310-262a-4852-89b4-4d1b1d6f5776 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1103.885157] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6f41e77-659a-4999-b3d4-4924ab4a74a2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.896884] env[70013]: DEBUG oslo_vmware.api [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1103.896884] env[70013]: value = "task-4231437" [ 1103.896884] env[70013]: _type = "Task" [ 1103.896884] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.912230] env[70013]: DEBUG oslo_vmware.api [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.918696] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4d36ed-04a7-4d79-a4e9-8b2d645c2b99 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.927018] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec62023-1e67-439b-b0bc-f2d1b45b37e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.962577] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b4b86f-6dc9-4acb-a280-9b6180f41291 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.976186] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a068ebaf-b35c-40c4-ade8-c2156201ab1e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.991145] env[70013]: DEBUG nova.compute.provider_tree [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.119255] env[70013]: DEBUG nova.compute.manager [req-05b68e2f-8159-4da8-9450-268d5ce5733d req-cdd5b0a2-0499-42b6-9e78-73c7cc2c6b23 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-vif-plugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1104.119255] env[70013]: DEBUG oslo_concurrency.lockutils [req-05b68e2f-8159-4da8-9450-268d5ce5733d req-cdd5b0a2-0499-42b6-9e78-73c7cc2c6b23 service nova] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.119255] env[70013]: DEBUG oslo_concurrency.lockutils [req-05b68e2f-8159-4da8-9450-268d5ce5733d req-cdd5b0a2-0499-42b6-9e78-73c7cc2c6b23 service nova] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.119255] env[70013]: DEBUG oslo_concurrency.lockutils [req-05b68e2f-8159-4da8-9450-268d5ce5733d req-cdd5b0a2-0499-42b6-9e78-73c7cc2c6b23 service nova] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.119255] env[70013]: DEBUG nova.compute.manager [req-05b68e2f-8159-4da8-9450-268d5ce5733d req-cdd5b0a2-0499-42b6-9e78-73c7cc2c6b23 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] No waiting events found dispatching network-vif-plugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1104.119255] env[70013]: WARNING nova.compute.manager [req-05b68e2f-8159-4da8-9450-268d5ce5733d req-cdd5b0a2-0499-42b6-9e78-73c7cc2c6b23 service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received unexpected event network-vif-plugged-3fc25179-89cf-42de-b6b6-f31806bcbce3 for instance with vm_state shelved_offloaded and task_state spawning. [ 1104.176761] env[70013]: DEBUG oslo_vmware.api [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231433, 'name': PowerOnVM_Task, 'duration_secs': 0.614501} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.177248] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.177499] env[70013]: INFO nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Took 7.36 seconds to spawn the instance on the hypervisor. [ 1104.177698] env[70013]: DEBUG nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1104.178525] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532b9334-435b-4870-9ddf-85491b32f9e1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.218031] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.218031] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1104.218031] env[70013]: DEBUG nova.network.neutron [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.318985] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231435, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648127} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.319876] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/8dd3e9c7-3a92-49e6-883c-cda647730f5e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1104.321999] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1104.321999] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3ed1f38-0312-4989-8cfb-6dac10c84c8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.333031] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1104.333031] env[70013]: value = "task-4231438" [ 1104.333031] env[70013]: _type = "Task" [ 1104.333031] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.342395] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231438, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.411745] env[70013]: DEBUG oslo_vmware.api [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.473859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.473859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.473859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.473859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.473859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.478321] env[70013]: INFO nova.compute.manager [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Terminating instance [ 1104.497018] env[70013]: DEBUG nova.scheduler.client.report [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1104.701438] env[70013]: INFO nova.compute.manager [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Took 20.37 seconds to build instance. [ 1104.845460] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231438, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074874} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.845904] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1104.847866] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cc832c-1b33-4c64-91db-e32c069ff891 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.880339] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/8dd3e9c7-3a92-49e6-883c-cda647730f5e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.880839] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9b14b18-d9a4-4e85-93d6-033d13370efb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.917714] env[70013]: DEBUG oslo_vmware.api [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.65538} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.919461] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.920020] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1104.920020] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1104.920157] env[70013]: INFO nova.compute.manager [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1104.920351] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1104.921325] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1104.921325] env[70013]: value = "task-4231439" [ 1104.921325] env[70013]: _type = "Task" [ 1104.921325] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.921325] env[70013]: DEBUG nova.compute.manager [-] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1104.921325] env[70013]: DEBUG nova.network.neutron [-] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1104.936989] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231439, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.981197] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "refresh_cache-456c4926-2ce4-4ba8-a8a6-ca50650e19f6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.981384] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquired lock "refresh_cache-456c4926-2ce4-4ba8-a8a6-ca50650e19f6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1104.981556] env[70013]: DEBUG nova.network.neutron [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.999149] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.581s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.002012] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.967s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.002742] env[70013]: DEBUG nova.objects.instance [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lazy-loading 'resources' on Instance uuid 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.156260] env[70013]: DEBUG nova.network.neutron [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fc25179-89", "ovs_interfaceid": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.206620] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9c066b41-ce50-4aad-91fc-2bbd5e936fc3 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.880s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.331495] env[70013]: DEBUG nova.compute.manager [req-65ab2f50-f996-4680-a04e-5d11eed8f5d9 req-9ea88a31-722e-439f-aa28-483d8f02d596 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Received event network-vif-deleted-f1b01fa4-dc57-40af-958b-ba24a089ab8d {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1105.331653] env[70013]: INFO nova.compute.manager [req-65ab2f50-f996-4680-a04e-5d11eed8f5d9 req-9ea88a31-722e-439f-aa28-483d8f02d596 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Neutron deleted interface f1b01fa4-dc57-40af-958b-ba24a089ab8d; detaching it from the instance and deleting it from the info cache [ 1105.331864] env[70013]: DEBUG nova.network.neutron [req-65ab2f50-f996-4680-a04e-5d11eed8f5d9 req-9ea88a31-722e-439f-aa28-483d8f02d596 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.438034] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231439, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.506455] env[70013]: DEBUG nova.network.neutron [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1105.515403] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e3b082f1-f5de-418f-9227-b9613e4e9ee6 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 32.807s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.516256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 7.045s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.516443] env[70013]: INFO nova.compute.manager [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Unshelving [ 1105.581375] env[70013]: DEBUG nova.network.neutron [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.661098] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1105.691980] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='267a7ff160efaadffadacfb711c571e4',container_format='bare',created_at=2025-05-19T11:39:06Z,direct_url=,disk_format='vmdk',id=16d5b38a-2afd-4319-88e0-3ebfc726db58,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1066076236-shelved',owner='0f5b0bd88f5046f098579c5d59e3f0c3',properties=ImageMetaProps,protected=,size=31670784,status='active',tags=,updated_at=2025-05-19T11:39:20Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1105.692293] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1105.692456] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1105.692643] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1105.692795] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1105.692982] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1105.695116] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1105.696571] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1105.696571] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1105.696571] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1105.696571] env[70013]: DEBUG nova.virt.hardware [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1105.696875] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fd0507-4832-4652-8c01-76330fc8a172 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.712226] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472280eb-31c2-42d7-965b-c66be1156dab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.730961] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:81:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3fc25179-89cf-42de-b6b6-f31806bcbce3', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1105.739914] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1105.739914] env[70013]: DEBUG nova.network.neutron [-] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.744435] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1105.744435] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ccf31b7-4b27-4bd1-85be-f9108ab47d48 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.765035] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f270d0c7-e1df-4792-930c-1a37e70b4416 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.776257] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e7fa2c4a-7c1d-4963-a2ad-51451976ae4b tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Suspending the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1105.777187] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-15976f2c-66c7-4598-bd44-56110a681bd6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.780312] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1105.780312] env[70013]: value = "task-4231440" [ 1105.780312] env[70013]: _type = "Task" [ 1105.780312] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.786833] env[70013]: DEBUG oslo_vmware.api [None req-e7fa2c4a-7c1d-4963-a2ad-51451976ae4b tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1105.786833] env[70013]: value = "task-4231441" [ 1105.786833] env[70013]: _type = "Task" [ 1105.786833] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.795533] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231440, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.801581] env[70013]: DEBUG oslo_vmware.api [None req-e7fa2c4a-7c1d-4963-a2ad-51451976ae4b tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231441, 'name': SuspendVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.836977] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ebbbeab4-5793-44c3-8e7f-30d4f97c526f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.847559] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e758c0b3-714d-4f60-b37b-a3892225565d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.887273] env[70013]: DEBUG nova.compute.manager [req-65ab2f50-f996-4680-a04e-5d11eed8f5d9 req-9ea88a31-722e-439f-aa28-483d8f02d596 service nova] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Detach interface failed, port_id=f1b01fa4-dc57-40af-958b-ba24a089ab8d, reason: Instance e3b37310-262a-4852-89b4-4d1b1d6f5776 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1105.941670] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231439, 'name': ReconfigVM_Task, 'duration_secs': 0.553847} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.941962] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/8dd3e9c7-3a92-49e6-883c-cda647730f5e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.942707] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6384611b-09d0-4a3d-8970-005c358cc8ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.954487] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1105.954487] env[70013]: value = "task-4231442" [ 1105.954487] env[70013]: _type = "Task" [ 1105.954487] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.964316] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231442, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.009939] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc09c080-f8b7-43d5-a0a6-7ee81ff3f91e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.020964] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238fc184-102b-42da-8326-a149a4220732 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.063395] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fd4421-ab9f-40e1-8027-fe892910d136 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.072607] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca22ada-1d7f-4bf0-978a-505eb2df6e40 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.090680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Releasing lock "refresh_cache-456c4926-2ce4-4ba8-a8a6-ca50650e19f6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1106.090680] env[70013]: DEBUG nova.compute.manager [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1106.090680] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1106.091258] env[70013]: DEBUG nova.compute.provider_tree [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.093342] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92906b93-cf5a-45b4-9fed-b9c0fafa5fab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.103015] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1106.103424] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21a851e0-5dae-448b-9068-25fe92f7a3e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.111127] env[70013]: DEBUG oslo_vmware.api [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1106.111127] env[70013]: value = "task-4231443" [ 1106.111127] env[70013]: _type = "Task" [ 1106.111127] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.124339] env[70013]: DEBUG oslo_vmware.api [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231443, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.186023] env[70013]: DEBUG nova.compute.manager [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1106.186023] env[70013]: DEBUG nova.compute.manager [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing instance network info cache due to event network-changed-3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1106.186480] env[70013]: DEBUG oslo_concurrency.lockutils [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] Acquiring lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.186480] env[70013]: DEBUG oslo_concurrency.lockutils [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] Acquired lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.186480] env[70013]: DEBUG nova.network.neutron [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Refreshing network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1106.244860] env[70013]: INFO nova.compute.manager [-] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Took 1.32 seconds to deallocate network for instance. [ 1106.292619] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231440, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.306656] env[70013]: DEBUG oslo_vmware.api [None req-e7fa2c4a-7c1d-4963-a2ad-51451976ae4b tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231441, 'name': SuspendVM_Task} progress is 58%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.330837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.330971] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1106.466575] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231442, 'name': Rename_Task, 'duration_secs': 0.171766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.466960] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1106.467303] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71ba2c6d-3ecc-4a01-ad4b-91982b5b2e95 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.476577] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1106.476577] env[70013]: value = "task-4231444" [ 1106.476577] env[70013]: _type = "Task" [ 1106.476577] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.486333] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.580546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.601028] env[70013]: DEBUG nova.scheduler.client.report [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1106.623318] env[70013]: DEBUG oslo_vmware.api [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231443, 'name': PowerOffVM_Task, 'duration_secs': 0.191343} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.624540] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1106.624875] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1106.625253] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2ca202d-ad5e-49bc-8838-cdb38cb55fa2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.654295] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1106.654842] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1106.655189] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Deleting the datastore file [datastore1] 456c4926-2ce4-4ba8-a8a6-ca50650e19f6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1106.655581] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bef6b09f-2caf-461b-a4fd-530b8660b889 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.663732] env[70013]: DEBUG oslo_vmware.api [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for the task: (returnval){ [ 1106.663732] env[70013]: value = "task-4231446" [ 1106.663732] env[70013]: _type = "Task" [ 1106.663732] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.673139] env[70013]: DEBUG oslo_vmware.api [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.754214] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.792016] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231440, 'name': CreateVM_Task, 'duration_secs': 0.526784} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.792652] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1106.796537] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.796964] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.797498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1106.797899] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c340d390-219c-4088-8871-ff62dfaee50f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.806269] env[70013]: DEBUG oslo_vmware.api [None req-e7fa2c4a-7c1d-4963-a2ad-51451976ae4b tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231441, 'name': SuspendVM_Task, 'duration_secs': 0.735931} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.808029] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e7fa2c4a-7c1d-4963-a2ad-51451976ae4b tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Suspended the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1106.808422] env[70013]: DEBUG nova.compute.manager [None req-e7fa2c4a-7c1d-4963-a2ad-51451976ae4b tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1106.808935] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1106.808935] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a8f129-52dc-ee5a-cec5-6962ac2549d8" [ 1106.808935] env[70013]: _type = "Task" [ 1106.808935] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.809944] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c90fee-d708-43dd-944d-e47adb230034 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.832302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1106.832714] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Processing image 16d5b38a-2afd-4319-88e0-3ebfc726db58 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1106.833241] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58/16d5b38a-2afd-4319-88e0-3ebfc726db58.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.833539] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58/16d5b38a-2afd-4319-88e0-3ebfc726db58.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.837048] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1106.837048] env[70013]: DEBUG nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1106.837540] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96144a15-4320-4c64-9416-8e3a19e48278 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.851530] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1106.851902] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1106.853157] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15dcc7b1-235d-46a8-a2a8-b822a9037240 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.860335] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1106.860335] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5209cd0b-6a17-3555-5af1-b7dc015eb325" [ 1106.860335] env[70013]: _type = "Task" [ 1106.860335] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.869992] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5209cd0b-6a17-3555-5af1-b7dc015eb325, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.978063] env[70013]: DEBUG nova.network.neutron [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updated VIF entry in instance network info cache for port 3fc25179-89cf-42de-b6b6-f31806bcbce3. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1106.978465] env[70013]: DEBUG nova.network.neutron [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [{"id": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "address": "fa:16:3e:96:81:a6", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fc25179-89", "ovs_interfaceid": "3fc25179-89cf-42de-b6b6-f31806bcbce3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.989952] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231444, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.104762] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.107221] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.301s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1107.108835] env[70013]: INFO nova.compute.claims [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1107.132533] env[70013]: INFO nova.scheduler.client.report [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Deleted allocations for instance 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6 [ 1107.178791] env[70013]: DEBUG oslo_vmware.api [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Task: {'id': task-4231446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143999} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.178791] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.179127] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.179332] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.179508] env[70013]: INFO nova.compute.manager [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Took 1.09 seconds to destroy the instance on the hypervisor. [ 1107.179920] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1107.180471] env[70013]: DEBUG nova.compute.manager [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1107.180577] env[70013]: DEBUG nova.network.neutron [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1107.203033] env[70013]: DEBUG nova.network.neutron [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1107.357060] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1107.372296] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Preparing fetch location {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1107.372552] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Fetch image to [datastore1] OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6/OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6.vmdk {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1107.372734] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Downloading stream optimized image 16d5b38a-2afd-4319-88e0-3ebfc726db58 to [datastore1] OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6/OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6.vmdk on the data store datastore1 as vApp {{(pid=70013) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1107.372926] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Downloading image file data 16d5b38a-2afd-4319-88e0-3ebfc726db58 to the ESX as VM named 'OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6' {{(pid=70013) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1107.456829] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1107.456829] env[70013]: value = "resgroup-9" [ 1107.456829] env[70013]: _type = "ResourcePool" [ 1107.456829] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1107.457154] env[70013]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-e34f1394-36b5-4303-a24f-1ca78eb40f05 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.479682] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease: (returnval){ [ 1107.479682] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a5bba-59eb-8e07-eeee-3e3c7936ba7e" [ 1107.479682] env[70013]: _type = "HttpNfcLease" [ 1107.479682] env[70013]: } obtained for vApp import into resource pool (val){ [ 1107.479682] env[70013]: value = "resgroup-9" [ 1107.479682] env[70013]: _type = "ResourcePool" [ 1107.479682] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1107.480075] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the lease: (returnval){ [ 1107.480075] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a5bba-59eb-8e07-eeee-3e3c7936ba7e" [ 1107.480075] env[70013]: _type = "HttpNfcLease" [ 1107.480075] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1107.487981] env[70013]: DEBUG oslo_concurrency.lockutils [req-fc723060-cf54-42cc-9d43-7ea82d4022e2 req-c6a8b911-f5a2-42f5-a9a6-2d6fcd00cadf service nova] Releasing lock "refresh_cache-472b8218-51c3-492e-96ff-5fa99df4cc4c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1107.494547] env[70013]: DEBUG oslo_vmware.api [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231444, 'name': PowerOnVM_Task, 'duration_secs': 0.531183} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.496178] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1107.496393] env[70013]: INFO nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Took 8.20 seconds to spawn the instance on the hypervisor. [ 1107.496573] env[70013]: DEBUG nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1107.496841] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1107.496841] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a5bba-59eb-8e07-eeee-3e3c7936ba7e" [ 1107.496841] env[70013]: _type = "HttpNfcLease" [ 1107.496841] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1107.497631] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eec4e34-8563-48d6-b6fc-fd99d82031f9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.641835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-39fc0e96-c803-4677-aed9-da9634742261 tempest-ServersTestBootFromVolume-1562632258 tempest-ServersTestBootFromVolume-1562632258-project-member] Lock "5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.713s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.704216] env[70013]: DEBUG nova.network.neutron [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.989906] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1107.989906] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a5bba-59eb-8e07-eeee-3e3c7936ba7e" [ 1107.989906] env[70013]: _type = "HttpNfcLease" [ 1107.989906] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1108.015902] env[70013]: INFO nova.compute.manager [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Took 22.21 seconds to build instance. [ 1108.209286] env[70013]: INFO nova.compute.manager [-] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Took 1.03 seconds to deallocate network for instance. [ 1108.479421] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc65583-7232-4f3c-9ec5-b634a05a445a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.491705] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c02b1a3-ac57-449f-871c-0c279275932e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.495911] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1108.495911] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a5bba-59eb-8e07-eeee-3e3c7936ba7e" [ 1108.495911] env[70013]: _type = "HttpNfcLease" [ 1108.495911] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1108.495911] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1108.495911] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a5bba-59eb-8e07-eeee-3e3c7936ba7e" [ 1108.495911] env[70013]: _type = "HttpNfcLease" [ 1108.495911] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1108.496453] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3aedb5-0405-4749-9e25-07028b7dd2fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.538362] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8084d991-a07b-4ba7-8e38-17eb3349b250 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.736s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.540436] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a27b23-8f89-4741-806c-6844bbeb8c5c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.546511] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521b0253-f2d1-1024-65d1-438cdb97acbe/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1108.546705] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating HTTP connection to write to file with size = 31670784 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521b0253-f2d1-1024-65d1-438cdb97acbe/disk-0.vmdk. {{(pid=70013) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1108.602056] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.602340] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.602554] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.602787] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.603021] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.609015] env[70013]: INFO nova.compute.manager [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Terminating instance [ 1108.620156] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6641259-b94d-4808-abf2-346f07203abe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.630499] env[70013]: DEBUG nova.compute.manager [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1108.630819] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1108.631173] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0d3d46eb-0216-4a39-8b98-0b7d27b4fdbf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.635735] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c716ecd-302b-4a47-8818-befeb14d724c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.644196] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1108.654149] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3129745-c958-44e0-adfc-210da110a977 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.656466] env[70013]: DEBUG nova.compute.provider_tree [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.723986] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.728798] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1108.729130] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1108.729360] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleting the datastore file [datastore1] bf138a0d-ea1f-4c93-9c7f-e0e3309597b2 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.729659] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-761c2594-df81-4239-9b77-157db3d58296 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.737118] env[70013]: DEBUG oslo_vmware.api [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1108.737118] env[70013]: value = "task-4231449" [ 1108.737118] env[70013]: _type = "Task" [ 1108.737118] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.747345] env[70013]: DEBUG oslo_vmware.api [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231449, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.116883] env[70013]: INFO nova.compute.manager [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Rescuing [ 1109.117273] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.117576] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.117824] env[70013]: DEBUG nova.network.neutron [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1109.163430] env[70013]: DEBUG nova.scheduler.client.report [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1109.252370] env[70013]: DEBUG oslo_vmware.api [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231449, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.261874} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.254393] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.254661] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1109.254915] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.255135] env[70013]: INFO nova.compute.manager [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1109.255414] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1109.255684] env[70013]: DEBUG nova.compute.manager [-] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1109.255875] env[70013]: DEBUG nova.network.neutron [-] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1109.671481] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.671481] env[70013]: DEBUG nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1109.680846] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.654s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.680846] env[70013]: DEBUG nova.objects.instance [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lazy-loading 'resources' on Instance uuid be86d96e-15be-4c39-b05d-59da0668c0d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.689103] env[70013]: DEBUG nova.compute.manager [req-dcc191a6-8f7b-4999-b2c5-925e08b2b558 req-9892e804-41ed-4d62-b662-21858761149c service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Received event network-vif-deleted-ba7e21fc-95aa-4d9c-b548-695aa34fbbb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1109.689358] env[70013]: INFO nova.compute.manager [req-dcc191a6-8f7b-4999-b2c5-925e08b2b558 req-9892e804-41ed-4d62-b662-21858761149c service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Neutron deleted interface ba7e21fc-95aa-4d9c-b548-695aa34fbbb7; detaching it from the instance and deleting it from the info cache [ 1109.689513] env[70013]: DEBUG nova.network.neutron [req-dcc191a6-8f7b-4999-b2c5-925e08b2b558 req-9892e804-41ed-4d62-b662-21858761149c service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.782349] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Completed reading data from the image iterator. {{(pid=70013) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1109.782767] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521b0253-f2d1-1024-65d1-438cdb97acbe/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1109.783717] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc4b667-0fd2-40a0-ab39-3ce70181cf96 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.795112] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521b0253-f2d1-1024-65d1-438cdb97acbe/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1109.795309] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521b0253-f2d1-1024-65d1-438cdb97acbe/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1109.795574] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-416326cc-e7f6-4147-9097-deba93d35d52 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.931659] env[70013]: DEBUG nova.network.neutron [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Updating instance_info_cache with network_info: [{"id": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "address": "fa:16:3e:43:93:9a", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2febcb-8e", "ovs_interfaceid": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.122484] env[70013]: DEBUG oslo_vmware.rw_handles [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521b0253-f2d1-1024-65d1-438cdb97acbe/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1110.122746] env[70013]: INFO nova.virt.vmwareapi.images [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Downloaded image file data 16d5b38a-2afd-4319-88e0-3ebfc726db58 [ 1110.123992] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c382bb5-0c80-4333-af4d-f23df7751b62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.149314] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82fe017d-c048-4300-959f-23c7a745cfb1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.153461] env[70013]: DEBUG nova.network.neutron [-] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.185250] env[70013]: DEBUG nova.compute.utils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1110.189664] env[70013]: DEBUG nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1110.189930] env[70013]: DEBUG nova.network.neutron [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1110.197975] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c1d4a48-01b1-4798-bb0e-2e4aefc28d5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.209213] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88332c4e-83ca-4578-9ead-882009a8bff7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.230556] env[70013]: DEBUG nova.policy [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1110.243524] env[70013]: DEBUG nova.compute.manager [req-dcc191a6-8f7b-4999-b2c5-925e08b2b558 req-9892e804-41ed-4d62-b662-21858761149c service nova] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Detach interface failed, port_id=ba7e21fc-95aa-4d9c-b548-695aa34fbbb7, reason: Instance bf138a0d-ea1f-4c93-9c7f-e0e3309597b2 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1110.247736] env[70013]: INFO nova.virt.vmwareapi.images [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] The imported VM was unregistered [ 1110.250393] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Caching image {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1110.250659] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating directory with path [datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.251475] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-10fb977c-3b25-44df-b1c4-740c6af05c2e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.285483] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created directory with path [datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.285711] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6/OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6.vmdk to [datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58/16d5b38a-2afd-4319-88e0-3ebfc726db58.vmdk. {{(pid=70013) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1110.285962] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-eaf561c4-0cff-4706-aa40-76995fdc9a8f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.295813] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1110.295813] env[70013]: value = "task-4231451" [ 1110.295813] env[70013]: _type = "Task" [ 1110.295813] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.303556] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.439017] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.609142] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20537fdb-aa78-4963-8624-8b6a45a85574 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.618175] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8b0eb3-4724-4e5b-9a82-bf5c2c999a1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.652861] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1c217c-af51-411a-bc3a-5386643b0325 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.656174] env[70013]: INFO nova.compute.manager [-] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Took 1.40 seconds to deallocate network for instance. [ 1110.665578] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980a44af-a926-444d-87d7-95dc29a13a91 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.684023] env[70013]: DEBUG nova.compute.provider_tree [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.690851] env[70013]: DEBUG nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1110.700169] env[70013]: DEBUG nova.network.neutron [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Successfully created port: 4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1110.806120] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.166683] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.190086] env[70013]: DEBUG nova.scheduler.client.report [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1111.307194] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.696482] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.017s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.700048] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.006s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.700227] env[70013]: DEBUG nova.objects.instance [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lazy-loading 'resources' on Instance uuid 8810bebc-448c-4e9b-9fbb-ce9e24611dc9 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.703010] env[70013]: DEBUG nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1111.723039] env[70013]: INFO nova.scheduler.client.report [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Deleted allocations for instance be86d96e-15be-4c39-b05d-59da0668c0d6 [ 1111.739020] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1111.739020] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1111.739020] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1111.739020] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1111.739020] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1111.739020] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1111.739640] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1111.740073] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1111.740493] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1111.742062] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1111.742062] env[70013]: DEBUG nova.virt.hardware [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1111.742062] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84698070-01c2-488f-99c1-90f6f4a8aecf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.752501] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2bfa3f8-4d25-48fb-8d8d-46fdcac5e9fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.807112] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.988712] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.988712] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f106c53-764f-4755-b8cb-b9fe7413d483 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.996025] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1111.996025] env[70013]: value = "task-4231452" [ 1111.996025] env[70013]: _type = "Task" [ 1111.996025] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.005289] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.241831] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2a03c0d5-a0ea-4709-a1de-122db3fae28b tempest-ServerPasswordTestJSON-1241276073 tempest-ServerPasswordTestJSON-1241276073-project-member] Lock "be86d96e-15be-4c39-b05d-59da0668c0d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.631s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.311397] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.485374] env[70013]: DEBUG nova.compute.manager [req-59d003ad-a5c8-4faf-a8d1-5ea717af08b0 req-588308c9-bfff-4b8b-afef-cd65a0bf093d service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-vif-plugged-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1112.485374] env[70013]: DEBUG oslo_concurrency.lockutils [req-59d003ad-a5c8-4faf-a8d1-5ea717af08b0 req-588308c9-bfff-4b8b-afef-cd65a0bf093d service nova] Acquiring lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.485374] env[70013]: DEBUG oslo_concurrency.lockutils [req-59d003ad-a5c8-4faf-a8d1-5ea717af08b0 req-588308c9-bfff-4b8b-afef-cd65a0bf093d service nova] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.485374] env[70013]: DEBUG oslo_concurrency.lockutils [req-59d003ad-a5c8-4faf-a8d1-5ea717af08b0 req-588308c9-bfff-4b8b-afef-cd65a0bf093d service nova] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.485374] env[70013]: DEBUG nova.compute.manager [req-59d003ad-a5c8-4faf-a8d1-5ea717af08b0 req-588308c9-bfff-4b8b-afef-cd65a0bf093d service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] No waiting events found dispatching network-vif-plugged-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1112.485374] env[70013]: WARNING nova.compute.manager [req-59d003ad-a5c8-4faf-a8d1-5ea717af08b0 req-588308c9-bfff-4b8b-afef-cd65a0bf093d service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received unexpected event network-vif-plugged-4d8f0295-7d55-4b26-803e-5f289211ed13 for instance with vm_state building and task_state spawning. [ 1112.508995] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231452, 'name': PowerOffVM_Task, 'duration_secs': 0.320219} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.508995] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1112.509104] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162aaf4b-d509-4387-89e9-12df049caf82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.537845] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d6a4f2-8352-48ef-89d5-a1263b11006f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.571160] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1112.571517] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7142b211-0ad5-4d80-8fcc-1cd779f1c31c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.579759] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1112.579759] env[70013]: value = "task-4231453" [ 1112.579759] env[70013]: _type = "Task" [ 1112.579759] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.592555] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1112.592555] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1112.592669] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.592743] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.592922] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1112.595832] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edef8821-f5ec-478e-bea4-146eb2acabe8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.606956] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1112.607259] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1112.607959] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd57c1e0-8259-4e48-898b-741145d20604 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.617608] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1112.617608] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521d8038-1d61-e311-26f1-d58f8bc16042" [ 1112.617608] env[70013]: _type = "Task" [ 1112.617608] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.629540] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521d8038-1d61-e311-26f1-d58f8bc16042, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.644817] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c91493-6fe2-414c-83f7-ff2145c8c55f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.653855] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa497ca4-cfd6-48b3-8ec2-cb9de18dfcf2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.685556] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee93009-eaf4-4181-a9f2-29aa2418c6f9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.696320] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738b1eb0-3a37-4384-83f3-cc91d8a538eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.711737] env[70013]: DEBUG nova.compute.provider_tree [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.811329] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.833171] env[70013]: DEBUG nova.network.neutron [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Successfully updated port: 4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1113.132119] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521d8038-1d61-e311-26f1-d58f8bc16042, 'name': SearchDatastore_Task, 'duration_secs': 0.023041} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.133088] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ad4ae34-5359-40f7-a272-db928e212a7b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.139598] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1113.139598] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bb6c90-6855-ce78-8d31-cf92ab70aa6a" [ 1113.139598] env[70013]: _type = "Task" [ 1113.139598] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.148975] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bb6c90-6855-ce78-8d31-cf92ab70aa6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.215261] env[70013]: DEBUG nova.scheduler.client.report [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1113.312357] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.335467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.335708] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.335904] env[70013]: DEBUG nova.network.neutron [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1113.498621] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "1230ca35-1512-4464-85fc-d3b4ab05eac1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.498895] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.499136] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "1230ca35-1512-4464-85fc-d3b4ab05eac1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.499325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.499493] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.501826] env[70013]: INFO nova.compute.manager [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Terminating instance [ 1113.658077] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bb6c90-6855-ce78-8d31-cf92ab70aa6a, 'name': SearchDatastore_Task, 'duration_secs': 0.016184} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.658077] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.658077] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. {{(pid=70013) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1113.658077] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-037f7918-15aa-494c-9d1d-4c5783f248b0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.659939] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1113.659939] env[70013]: value = "task-4231454" [ 1113.659939] env[70013]: _type = "Task" [ 1113.659939] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.669198] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231454, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.723653] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.024s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.727558] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.110s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.727766] env[70013]: DEBUG nova.objects.instance [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1113.754500] env[70013]: INFO nova.scheduler.client.report [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Deleted allocations for instance 8810bebc-448c-4e9b-9fbb-ce9e24611dc9 [ 1113.817915] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231451, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.215101} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.818097] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6/OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6.vmdk to [datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58/16d5b38a-2afd-4319-88e0-3ebfc726db58.vmdk. [ 1113.818618] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Cleaning up location [datastore1] OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1113.818618] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_e9607038-fb0c-44b9-92e4-f82fa4a877f6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.819055] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cac02763-ed58-4deb-bbd1-aff5e2752d28 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.827900] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1113.827900] env[70013]: value = "task-4231455" [ 1113.827900] env[70013]: _type = "Task" [ 1113.827900] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.838156] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.880066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "addf8ba9-142c-4458-8586-b06e52118d71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.880360] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "addf8ba9-142c-4458-8586-b06e52118d71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.880553] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "addf8ba9-142c-4458-8586-b06e52118d71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.880739] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "addf8ba9-142c-4458-8586-b06e52118d71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.880917] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "addf8ba9-142c-4458-8586-b06e52118d71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.886252] env[70013]: INFO nova.compute.manager [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Terminating instance [ 1113.900842] env[70013]: DEBUG nova.network.neutron [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1114.007307] env[70013]: DEBUG nova.compute.manager [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1114.007550] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1114.008910] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bdf972-ad66-4104-aaf2-9e504e03c653 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.020012] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.020404] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e7288ce-83fa-4019-b449-affe19417738 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.031945] env[70013]: DEBUG oslo_vmware.api [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1114.031945] env[70013]: value = "task-4231456" [ 1114.031945] env[70013]: _type = "Task" [ 1114.031945] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.043917] env[70013]: DEBUG oslo_vmware.api [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231456, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.172598] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231454, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.266270] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f525d498-b991-45f2-b892-924426ec8c4a tempest-AttachInterfacesUnderV243Test-1270413728 tempest-AttachInterfacesUnderV243Test-1270413728-project-member] Lock "8810bebc-448c-4e9b-9fbb-ce9e24611dc9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.684s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.338428] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.057313} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.338702] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.338864] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58/16d5b38a-2afd-4319-88e0-3ebfc726db58.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1114.339132] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58/16d5b38a-2afd-4319-88e0-3ebfc726db58.vmdk to [datastore1] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1114.339393] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a48d69d6-f0db-4f09-8125-e43d5f6374ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.346869] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1114.346869] env[70013]: value = "task-4231457" [ 1114.346869] env[70013]: _type = "Task" [ 1114.346869] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.355537] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.390700] env[70013]: DEBUG nova.compute.manager [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1114.390985] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1114.391929] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a3896b-61eb-4e66-8c6e-b2d818adcc5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.399643] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.399917] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21146e18-234a-41ed-8f00-f0a5f54aa3a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.406890] env[70013]: DEBUG oslo_vmware.api [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1114.406890] env[70013]: value = "task-4231458" [ 1114.406890] env[70013]: _type = "Task" [ 1114.406890] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.418026] env[70013]: DEBUG oslo_vmware.api [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231458, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.543044] env[70013]: DEBUG oslo_vmware.api [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231456, 'name': PowerOffVM_Task, 'duration_secs': 0.310535} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.543340] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.543510] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1114.544136] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc686850-b910-44ac-ab5f-9305872a6c5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.564600] env[70013]: DEBUG nova.network.neutron [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.620540] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1114.620763] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1114.620948] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleting the datastore file [datastore2] 1230ca35-1512-4464-85fc-d3b4ab05eac1 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.621251] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ced8a5fa-5007-4583-9d50-57392ebbfb5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.629853] env[70013]: DEBUG oslo_vmware.api [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1114.629853] env[70013]: value = "task-4231460" [ 1114.629853] env[70013]: _type = "Task" [ 1114.629853] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.640096] env[70013]: DEBUG oslo_vmware.api [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231460, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.671791] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231454, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527882} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.672130] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk. [ 1114.672985] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954b6daa-ba61-4857-892c-dc1ca27541c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.701860] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.702253] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-258bf983-f3c5-49bc-8a2f-eb689fda1059 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.718849] env[70013]: DEBUG nova.compute.manager [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1114.719060] env[70013]: DEBUG nova.compute.manager [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing instance network info cache due to event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1114.719264] env[70013]: DEBUG oslo_concurrency.lockutils [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.556805] env[70013]: DEBUG oslo_concurrency.lockutils [None req-20599dc8-bacb-4d92-a073-59f82b06f087 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.829s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1115.558171] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.558600] env[70013]: DEBUG nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Instance network_info: |[{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1115.558987] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1115.558987] env[70013]: value = "task-4231461" [ 1115.558987] env[70013]: _type = "Task" [ 1115.558987] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.559681] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.979s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.559681] env[70013]: DEBUG nova.objects.instance [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lazy-loading 'pci_requests' on Instance uuid 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.567165] env[70013]: DEBUG oslo_concurrency.lockutils [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.567299] env[70013]: DEBUG nova.network.neutron [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1115.568665] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:cb:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d8f0295-7d55-4b26-803e-5f289211ed13', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.576401] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1115.582932] env[70013]: DEBUG nova.objects.instance [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lazy-loading 'numa_topology' on Instance uuid 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.584626] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1115.589216] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c96847b-1fa4-4747-8a01-1954963a273a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.612069] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231457, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.613129] env[70013]: DEBUG oslo_vmware.api [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231458, 'name': PowerOffVM_Task, 'duration_secs': 0.23323} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.617679] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1115.617679] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1115.617679] env[70013]: DEBUG oslo_vmware.api [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231460, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.304924} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.620510] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16a12ef0-a8b0-4006-aea6-78094eb5acef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.622313] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.622518] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1115.622695] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1115.623658] env[70013]: INFO nova.compute.manager [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1115.623658] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1115.623658] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231461, 'name': ReconfigVM_Task, 'duration_secs': 0.642877} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.625081] env[70013]: DEBUG nova.compute.manager [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1115.625166] env[70013]: DEBUG nova.network.neutron [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1115.626879] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e/0ddeef65-a14c-4065-97bb-58b607968d19-rescue.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.627467] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.627467] env[70013]: value = "task-4231462" [ 1115.627467] env[70013]: _type = "Task" [ 1115.627467] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.628477] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92832bcf-470e-4e17-bf21-bbd529d10a62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.642827] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231462, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.670537] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e444d7a7-7abb-4b2c-91cd-f7922a52982d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.688760] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1115.688760] env[70013]: value = "task-4231464" [ 1115.688760] env[70013]: _type = "Task" [ 1115.688760] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.700091] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231464, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.818317] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1115.818543] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1115.818732] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Deleting the datastore file [datastore2] addf8ba9-142c-4458-8586-b06e52118d71 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.819024] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a3fb4c1-b8eb-445c-a4c4-993ae9bf2fb8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.827586] env[70013]: DEBUG oslo_vmware.api [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for the task: (returnval){ [ 1115.827586] env[70013]: value = "task-4231465" [ 1115.827586] env[70013]: _type = "Task" [ 1115.827586] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.837321] env[70013]: DEBUG oslo_vmware.api [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231465, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.069803] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231457, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.088907] env[70013]: INFO nova.compute.claims [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.147135] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231462, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.201586] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231464, 'name': ReconfigVM_Task, 'duration_secs': 0.293288} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.201887] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1116.202200] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77d52a1b-baf0-4a9d-b3d7-2df106d61d06 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.209586] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1116.209586] env[70013]: value = "task-4231466" [ 1116.209586] env[70013]: _type = "Task" [ 1116.209586] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.218939] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231466, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.339318] env[70013]: DEBUG oslo_vmware.api [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Task: {'id': task-4231465, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.333295} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.339598] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1116.339782] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1116.339990] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1116.340293] env[70013]: INFO nova.compute.manager [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Took 1.95 seconds to destroy the instance on the hypervisor. [ 1116.340552] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1116.340760] env[70013]: DEBUG nova.compute.manager [-] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1116.340857] env[70013]: DEBUG nova.network.neutron [-] [instance: addf8ba9-142c-4458-8586-b06e52118d71] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1116.572774] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231457, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.577130] env[70013]: DEBUG nova.network.neutron [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updated VIF entry in instance network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1116.577503] env[70013]: DEBUG nova.network.neutron [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.645908] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231462, 'name': CreateVM_Task, 'duration_secs': 0.584196} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.646155] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1116.648752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.648752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.648752] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1116.648752] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-821d6024-cd5e-49b3-8496-6756056f836d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.657415] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1116.657415] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5246febd-e131-2f13-5ce3-6acf33273aa5" [ 1116.657415] env[70013]: _type = "Task" [ 1116.657415] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.668794] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5246febd-e131-2f13-5ce3-6acf33273aa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.729623] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231466, 'name': PowerOnVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.780806] env[70013]: DEBUG nova.compute.manager [req-275f2f84-17ea-4267-88ea-c49911e03478 req-b68450bd-cec2-4414-b620-374dc4c84286 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Received event network-vif-deleted-3d1c82dc-c975-4156-8d68-368f68ec3f87 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1116.780967] env[70013]: INFO nova.compute.manager [req-275f2f84-17ea-4267-88ea-c49911e03478 req-b68450bd-cec2-4414-b620-374dc4c84286 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Neutron deleted interface 3d1c82dc-c975-4156-8d68-368f68ec3f87; detaching it from the instance and deleting it from the info cache [ 1116.781158] env[70013]: DEBUG nova.network.neutron [req-275f2f84-17ea-4267-88ea-c49911e03478 req-b68450bd-cec2-4414-b620-374dc4c84286 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.976913] env[70013]: DEBUG nova.network.neutron [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.033307] env[70013]: DEBUG nova.compute.manager [req-b77d4d7d-0d21-4c23-be01-0e8b523a4cc7 req-55eb1b6b-c03b-44fa-8947-75cfe68cc8cf service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Received event network-vif-deleted-4151335a-4b19-45bd-bf10-69e4960fe456 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1117.033528] env[70013]: INFO nova.compute.manager [req-b77d4d7d-0d21-4c23-be01-0e8b523a4cc7 req-55eb1b6b-c03b-44fa-8947-75cfe68cc8cf service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Neutron deleted interface 4151335a-4b19-45bd-bf10-69e4960fe456; detaching it from the instance and deleting it from the info cache [ 1117.033701] env[70013]: DEBUG nova.network.neutron [req-b77d4d7d-0d21-4c23-be01-0e8b523a4cc7 req-55eb1b6b-c03b-44fa-8947-75cfe68cc8cf service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.071540] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231457, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.552174} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.071960] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/16d5b38a-2afd-4319-88e0-3ebfc726db58/16d5b38a-2afd-4319-88e0-3ebfc726db58.vmdk to [datastore1] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1117.073528] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652dd89e-9787-4ae3-9c01-baa8673f306e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.090654] env[70013]: DEBUG oslo_concurrency.lockutils [req-b3f51574-dfd1-42f9-95e6-f4c10387fb1e req-c1920808-b815-4131-ace4-ba628db61a93 service nova] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.099960] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.103419] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc951dcd-ee7e-4e39-b1c0-867d35e367a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.126554] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1117.126554] env[70013]: value = "task-4231467" [ 1117.126554] env[70013]: _type = "Task" [ 1117.126554] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.137385] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231467, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.169293] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5246febd-e131-2f13-5ce3-6acf33273aa5, 'name': SearchDatastore_Task, 'duration_secs': 0.093939} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.172947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.173896] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1117.174119] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.175039] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.175039] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.175417] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2776b4b-4a02-48de-9927-60dc567a70eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.195489] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.195761] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1117.197720] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cdda47c-64f1-4a29-9b0c-8fde5bffd5d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.203756] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1117.203756] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d7ba5f-d44c-9af6-0f09-33dea01825f0" [ 1117.203756] env[70013]: _type = "Task" [ 1117.203756] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.218472] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d7ba5f-d44c-9af6-0f09-33dea01825f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.224302] env[70013]: DEBUG oslo_vmware.api [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231466, 'name': PowerOnVM_Task, 'duration_secs': 0.604208} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.224722] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1117.229937] env[70013]: DEBUG nova.compute.manager [None req-8d8d9951-85de-4049-ab5d-9351a2686176 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1117.230845] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c89253f-0f57-42db-bbb6-7a3f74c61476 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.290038] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45045c21-5529-410f-aa3d-a80957a6c806 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.300862] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8adef948-7d8a-4685-a495-d3c19b5a47aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.336569] env[70013]: DEBUG nova.compute.manager [req-275f2f84-17ea-4267-88ea-c49911e03478 req-b68450bd-cec2-4414-b620-374dc4c84286 service nova] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Detach interface failed, port_id=3d1c82dc-c975-4156-8d68-368f68ec3f87, reason: Instance 1230ca35-1512-4464-85fc-d3b4ab05eac1 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1117.478971] env[70013]: INFO nova.compute.manager [-] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Took 1.85 seconds to deallocate network for instance. [ 1117.515160] env[70013]: DEBUG nova.network.neutron [-] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.520591] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc069768-7de3-4835-876b-7dbcf86b1a0d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.530216] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc724a2f-119c-4c60-abad-021431e92893 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.578035] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfa81a7a-279c-43c4-af57-208ff96287e9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.580784] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfd43c2-dc3a-4279-8815-dc8eec8d0383 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.589894] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80b837f-69bf-461b-974a-0254a7a01c61 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.596860] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43adbf2b-5700-4e5e-9e25-506dac5c1751 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.617992] env[70013]: DEBUG nova.compute.provider_tree [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.627505] env[70013]: DEBUG nova.compute.manager [req-b77d4d7d-0d21-4c23-be01-0e8b523a4cc7 req-55eb1b6b-c03b-44fa-8947-75cfe68cc8cf service nova] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Detach interface failed, port_id=4151335a-4b19-45bd-bf10-69e4960fe456, reason: Instance addf8ba9-142c-4458-8586-b06e52118d71 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1117.638800] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231467, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.715975] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d7ba5f-d44c-9af6-0f09-33dea01825f0, 'name': SearchDatastore_Task, 'duration_secs': 0.06583} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.716201] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1cc961d-2673-4c5a-ac24-2110eed2d0b6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.722962] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1117.722962] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5231a71d-080a-6215-7e19-f7718bc79379" [ 1117.722962] env[70013]: _type = "Task" [ 1117.722962] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.731924] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5231a71d-080a-6215-7e19-f7718bc79379, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.990643] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.018614] env[70013]: INFO nova.compute.manager [-] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Took 1.68 seconds to deallocate network for instance. [ 1118.134971] env[70013]: DEBUG nova.scheduler.client.report [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1118.144902] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231467, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.234424] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5231a71d-080a-6215-7e19-f7718bc79379, 'name': SearchDatastore_Task, 'duration_secs': 0.036905} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.234695] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.234954] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 38738d80-25ca-4a17-81f1-e0a7bc8c4a44/38738d80-25ca-4a17-81f1-e0a7bc8c4a44.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1118.235366] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c8defe8-8e54-402a-98a4-e477220b6b0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.243406] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1118.243406] env[70013]: value = "task-4231468" [ 1118.243406] env[70013]: _type = "Task" [ 1118.243406] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.252135] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231468, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.530590] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.638352] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231467, 'name': ReconfigVM_Task, 'duration_secs': 1.241413} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.639520] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.080s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1118.641797] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 472b8218-51c3-492e-96ff-5fa99df4cc4c/472b8218-51c3-492e-96ff-5fa99df4cc4c.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.644274] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encrypted': False, 'encryption_secret_uuid': None, 'device_type': 'disk', 'boot_index': 0, 'encryption_options': None, 'disk_bus': None, 'size': 0, 'encryption_format': None, 'device_name': '/dev/sda', 'image_id': '0ddeef65-a14c-4065-97bb-58b607968d19'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '06422efa-d203-436d-97c1-e3d13b343c13', 'guest_format': None, 'device_type': None, 'mount_device': '/dev/sdb', 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837209', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'name': 'volume-5a787b14-e921-4a07-a8c6-06fd861030f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '472b8218-51c3-492e-96ff-5fa99df4cc4c', 'attached_at': '', 'detached_at': '', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'serial': '5a787b14-e921-4a07-a8c6-06fd861030f1'}, 'disk_bus': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=70013) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1118.644274] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1118.644274] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837209', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'name': 'volume-5a787b14-e921-4a07-a8c6-06fd861030f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '472b8218-51c3-492e-96ff-5fa99df4cc4c', 'attached_at': '', 'detached_at': '', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'serial': '5a787b14-e921-4a07-a8c6-06fd861030f1'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1118.644274] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.890s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.645535] env[70013]: DEBUG nova.objects.instance [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'resources' on Instance uuid e3b37310-262a-4852-89b4-4d1b1d6f5776 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.646613] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd46336c-5fbf-46bf-8a26-837523b7e09f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.664330] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32d921a-db50-4c1b-8ca9-f6f5909e6d59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.702587] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] volume-5a787b14-e921-4a07-a8c6-06fd861030f1/volume-5a787b14-e921-4a07-a8c6-06fd861030f1.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.704807] env[70013]: INFO nova.network.neutron [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating port d8aacd8d-2fc2-477b-9d82-e101e20b0132 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1118.709644] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e486e8ef-126d-4927-b5a9-40c1bb961803 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.735078] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1118.735078] env[70013]: value = "task-4231469" [ 1118.735078] env[70013]: _type = "Task" [ 1118.735078] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.742360] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231469, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.753306] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231468, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.894124] env[70013]: INFO nova.compute.manager [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Unrescuing [ 1118.894124] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.894124] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquired lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1118.894124] env[70013]: DEBUG nova.network.neutron [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1119.252390] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231469, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.258611] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231468, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.783922} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.258867] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 38738d80-25ca-4a17-81f1-e0a7bc8c4a44/38738d80-25ca-4a17-81f1-e0a7bc8c4a44.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1119.259096] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1119.259354] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60ae4625-4683-4b20-aead-32050c02d41d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.266768] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1119.266768] env[70013]: value = "task-4231470" [ 1119.266768] env[70013]: _type = "Task" [ 1119.266768] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.280099] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.548651] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eaff2ed-8710-4be2-acd4-cf83059f4e7d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.557798] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6b2008-11d0-46d4-988a-02f3063129fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.595920] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de83478-446d-4703-9406-df995dfa8275 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.604919] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9b129c-7b3a-4342-b1aa-0488f95fde4f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.621247] env[70013]: DEBUG nova.compute.provider_tree [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.743757] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231469, 'name': ReconfigVM_Task, 'duration_secs': 0.569506} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.744293] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfigured VM instance instance-0000003d to attach disk [datastore1] volume-5a787b14-e921-4a07-a8c6-06fd861030f1/volume-5a787b14-e921-4a07-a8c6-06fd861030f1.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.750158] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e843b6b-fde4-48de-8ac2-a6334c7fd979 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.761250] env[70013]: DEBUG nova.network.neutron [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Updating instance_info_cache with network_info: [{"id": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "address": "fa:16:3e:43:93:9a", "network": {"id": "d2ef0f03-2dcd-4591-b08a-e1f19a1e799b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-658297555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3c6094eeb5ca48dfbfbb10bc71feac67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2febcb-8e", "ovs_interfaceid": "1d2febcb-8ea3-4126-8de7-7bbe2f7162ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.776124] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1119.776124] env[70013]: value = "task-4231471" [ 1119.776124] env[70013]: _type = "Task" [ 1119.776124] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.785669] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.422063} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.786872] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1119.787754] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f290832-56ea-4e01-8ce6-6bd791b8360b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.796092] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231471, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.819324] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 38738d80-25ca-4a17-81f1-e0a7bc8c4a44/38738d80-25ca-4a17-81f1-e0a7bc8c4a44.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.820144] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b09d81fa-ec4c-4cb3-bb09-5922f7fd58b4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.840330] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1119.840330] env[70013]: value = "task-4231472" [ 1119.840330] env[70013]: _type = "Task" [ 1119.840330] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.849623] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231472, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.124887] env[70013]: DEBUG nova.scheduler.client.report [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1120.264069] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Releasing lock "refresh_cache-8dd3e9c7-3a92-49e6-883c-cda647730f5e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1120.264793] env[70013]: DEBUG nova.objects.instance [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lazy-loading 'flavor' on Instance uuid 8dd3e9c7-3a92-49e6-883c-cda647730f5e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.288377] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231471, 'name': ReconfigVM_Task, 'duration_secs': 0.35541} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.289466] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837209', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'name': 'volume-5a787b14-e921-4a07-a8c6-06fd861030f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '472b8218-51c3-492e-96ff-5fa99df4cc4c', 'attached_at': '', 'detached_at': '', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'serial': '5a787b14-e921-4a07-a8c6-06fd861030f1'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1120.289466] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d41a7332-df0f-47cc-ac3d-7a805edf17d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.296432] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1120.296432] env[70013]: value = "task-4231473" [ 1120.296432] env[70013]: _type = "Task" [ 1120.296432] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.305380] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231473, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.354407] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231472, 'name': ReconfigVM_Task, 'duration_secs': 0.30218} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.354710] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 38738d80-25ca-4a17-81f1-e0a7bc8c4a44/38738d80-25ca-4a17-81f1-e0a7bc8c4a44.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.355389] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-566deee1-0d34-4d9c-895b-27b0ddca4061 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.363341] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1120.363341] env[70013]: value = "task-4231474" [ 1120.363341] env[70013]: _type = "Task" [ 1120.363341] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.376772] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231474, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.631561] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.639997] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.280s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.644371] env[70013]: INFO nova.compute.claims [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1120.665719] env[70013]: DEBUG nova.compute.manager [req-be8139bc-9022-47e1-9148-df522b8b9801 req-4ac2d14e-d28c-4f75-b7f6-b66160e2ff4a service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received event network-vif-plugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1120.666052] env[70013]: DEBUG oslo_concurrency.lockutils [req-be8139bc-9022-47e1-9148-df522b8b9801 req-4ac2d14e-d28c-4f75-b7f6-b66160e2ff4a service nova] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.666337] env[70013]: DEBUG oslo_concurrency.lockutils [req-be8139bc-9022-47e1-9148-df522b8b9801 req-4ac2d14e-d28c-4f75-b7f6-b66160e2ff4a service nova] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.666554] env[70013]: DEBUG oslo_concurrency.lockutils [req-be8139bc-9022-47e1-9148-df522b8b9801 req-4ac2d14e-d28c-4f75-b7f6-b66160e2ff4a service nova] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.666749] env[70013]: DEBUG nova.compute.manager [req-be8139bc-9022-47e1-9148-df522b8b9801 req-4ac2d14e-d28c-4f75-b7f6-b66160e2ff4a service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] No waiting events found dispatching network-vif-plugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1120.666971] env[70013]: WARNING nova.compute.manager [req-be8139bc-9022-47e1-9148-df522b8b9801 req-4ac2d14e-d28c-4f75-b7f6-b66160e2ff4a service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received unexpected event network-vif-plugged-d8aacd8d-2fc2-477b-9d82-e101e20b0132 for instance with vm_state shelved_offloaded and task_state spawning. [ 1120.671778] env[70013]: INFO nova.scheduler.client.report [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted allocations for instance e3b37310-262a-4852-89b4-4d1b1d6f5776 [ 1120.765319] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.765319] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1120.765319] env[70013]: DEBUG nova.network.neutron [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1120.776659] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda048c5-c08b-4fa1-b7b7-465630d947f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.804833] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1120.810055] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bad40fd3-361c-4b49-bfda-7618b8763b2b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.820361] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231473, 'name': Rename_Task, 'duration_secs': 0.188874} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.821487] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1120.821831] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1120.821831] env[70013]: value = "task-4231475" [ 1120.821831] env[70013]: _type = "Task" [ 1120.821831] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.822290] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dba73b0b-5e42-40aa-a647-237da897c944 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.834566] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.836056] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1120.836056] env[70013]: value = "task-4231476" [ 1120.836056] env[70013]: _type = "Task" [ 1120.836056] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.844801] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231476, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.874743] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231474, 'name': Rename_Task, 'duration_secs': 0.151745} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.875067] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1120.875756] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f69ca40-0bff-4496-9c27-c9fd33ea9ebe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.884337] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1120.884337] env[70013]: value = "task-4231477" [ 1120.884337] env[70013]: _type = "Task" [ 1120.884337] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.894691] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.181715] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aef764b8-21bc-4144-8c4b-42c254c0fc3e tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e3b37310-262a-4852-89b4-4d1b1d6f5776" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.440s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.337899] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231475, 'name': PowerOffVM_Task, 'duration_secs': 0.28004} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.341953] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1121.349077] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1121.349502] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29403a6d-eba4-4398-96f7-3caff133d0de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.386757] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231476, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.393648] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1121.393648] env[70013]: value = "task-4231478" [ 1121.393648] env[70013]: _type = "Task" [ 1121.393648] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.408108] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231477, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.415344] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231478, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.547489] env[70013]: DEBUG nova.network.neutron [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.849163] env[70013]: DEBUG oslo_vmware.api [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231476, 'name': PowerOnVM_Task, 'duration_secs': 0.79594} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.852519] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.902793] env[70013]: DEBUG oslo_vmware.api [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231477, 'name': PowerOnVM_Task, 'duration_secs': 0.533797} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.906639] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.906866] env[70013]: INFO nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Took 10.20 seconds to spawn the instance on the hypervisor. [ 1121.907062] env[70013]: DEBUG nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1121.907868] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0603bb-c477-4ea8-9eb5-8c738b3209eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.922739] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231478, 'name': ReconfigVM_Task, 'duration_secs': 0.317464} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.923341] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1121.923592] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1121.927036] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba969afa-b4e5-4911-8f39-f94c6d865823 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.938997] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1121.938997] env[70013]: value = "task-4231479" [ 1121.938997] env[70013]: _type = "Task" [ 1121.938997] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.960749] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.996058] env[70013]: DEBUG nova.compute.manager [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1121.997109] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e788024e-f4f2-4231-ae01-26885df2dfcf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.051068] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1122.077781] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3073e1b6aaca175d2d735cecc9d8dc86',container_format='bare',created_at=2025-05-19T11:39:30Z,direct_url=,disk_format='vmdk',id=ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-162683803-shelved',owner='68ad346a8e3a465da70d3a7de825ac6d',properties=ImageMetaProps,protected=,size=31662592,status='active',tags=,updated_at=2025-05-19T11:39:47Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1122.077781] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.077781] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1122.078026] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.078075] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1122.078252] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1122.078929] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1122.078929] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1122.079081] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1122.079229] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1122.079478] env[70013]: DEBUG nova.virt.hardware [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1122.080766] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f0efc8-63ac-4946-8f5b-713413c34883 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.085501] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08c0907-d3a1-4dc4-b0a0-480b27588489 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.099098] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e4808c-b0a5-4f9f-8a1b-8d042c3f258a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.104840] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7d9f1d-7156-462f-bbc3-f412426ba5b0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.119945] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:be:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8aacd8d-2fc2-477b-9d82-e101e20b0132', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1122.127688] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1122.155829] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1122.156694] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ef54e3d-71a2-4634-84ce-3fe5b239fdda {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.172043] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904300f7-0364-409c-91ce-a1c9666b12d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.181876] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49604f2-0d56-4488-bd72-0243b04b6ca0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.187037] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1122.187037] env[70013]: value = "task-4231480" [ 1122.187037] env[70013]: _type = "Task" [ 1122.187037] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.199938] env[70013]: DEBUG nova.compute.provider_tree [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.207584] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231480, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.440239] env[70013]: INFO nova.compute.manager [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Took 25.65 seconds to build instance. [ 1122.464595] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231479, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.526428] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6f57392a-4d1e-4e84-9d36-a9bdc4f56f8a tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 37.948s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.654386] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "499bb115-146c-48dc-9447-3c01c9549a73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.654636] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "499bb115-146c-48dc-9447-3c01c9549a73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.697475] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231480, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.700029] env[70013]: DEBUG nova.compute.manager [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received event network-changed-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1122.700344] env[70013]: DEBUG nova.compute.manager [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Refreshing instance network info cache due to event network-changed-d8aacd8d-2fc2-477b-9d82-e101e20b0132. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1122.700503] env[70013]: DEBUG oslo_concurrency.lockutils [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] Acquiring lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.700655] env[70013]: DEBUG oslo_concurrency.lockutils [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] Acquired lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1122.700844] env[70013]: DEBUG nova.network.neutron [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Refreshing network info cache for port d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1122.703217] env[70013]: DEBUG nova.scheduler.client.report [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1122.948152] env[70013]: DEBUG oslo_concurrency.lockutils [None req-830081bc-78c5-403f-b76d-3c5cb203352e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.170s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.958687] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231479, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.980323] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "e11550bf-3284-4ae6-9aa9-34923da589ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.980616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.161026] env[70013]: DEBUG nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1123.202690] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231480, 'name': CreateVM_Task, 'duration_secs': 0.844144} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.209259] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1123.210780] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.213031] env[70013]: DEBUG nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1123.219496] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.219496] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.219496] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1123.221061] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.497s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.221300] env[70013]: DEBUG nova.objects.instance [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lazy-loading 'resources' on Instance uuid 456c4926-2ce4-4ba8-a8a6-ca50650e19f6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.222973] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-209c8d86-4364-43cb-946d-0d3a5d2d146c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.230346] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1123.230346] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52445d21-cd65-9c80-9af6-c8ecdfcc2935" [ 1123.230346] env[70013]: _type = "Task" [ 1123.230346] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.242825] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52445d21-cd65-9c80-9af6-c8ecdfcc2935, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.430478] env[70013]: DEBUG nova.network.neutron [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updated VIF entry in instance network info cache for port d8aacd8d-2fc2-477b-9d82-e101e20b0132. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1123.430856] env[70013]: DEBUG nova.network.neutron [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.458258] env[70013]: DEBUG oslo_vmware.api [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231479, 'name': PowerOnVM_Task, 'duration_secs': 1.229161} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.458543] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1123.458775] env[70013]: DEBUG nova.compute.manager [None req-ab2dc432-f5ab-41d9-9e28-86c9f9f42f6a tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1123.459593] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69e1605-161a-4f7a-bd8a-6dfcc5c2e4d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.484052] env[70013]: DEBUG nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1123.652690] env[70013]: DEBUG nova.compute.manager [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1123.652690] env[70013]: DEBUG nova.compute.manager [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing instance network info cache due to event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1123.652917] env[70013]: DEBUG oslo_concurrency.lockutils [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.652994] env[70013]: DEBUG oslo_concurrency.lockutils [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.654026] env[70013]: DEBUG nova.network.neutron [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1123.686334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.722903] env[70013]: DEBUG nova.compute.utils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1123.726026] env[70013]: DEBUG nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1123.726026] env[70013]: DEBUG nova.network.neutron [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1123.748905] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1123.749350] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Processing image ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1123.749754] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.750344] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.750517] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1123.751289] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b46a876-a70c-422a-92fb-a6cc699c1b68 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.768226] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1123.768436] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1123.769396] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b4f642b-de13-4ed3-8952-95ec04f5b23b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.781526] env[70013]: DEBUG nova.policy [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38f449100e964d92b72b2ab8d14c550d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '77d07f4b4bd24e40883d1a381ec8e002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1123.786014] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1123.786014] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523cb8c3-f78d-37f3-f7a3-65459bb92b57" [ 1123.786014] env[70013]: _type = "Task" [ 1123.786014] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.798899] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Preparing fetch location {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1123.799244] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Fetch image to [datastore1] OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632/OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632.vmdk {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1123.799421] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Downloading stream optimized image ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 to [datastore1] OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632/OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632.vmdk on the data store datastore1 as vApp {{(pid=70013) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1123.799594] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Downloading image file data ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 to the ESX as VM named 'OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632' {{(pid=70013) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1123.893540] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1123.893540] env[70013]: value = "resgroup-9" [ 1123.893540] env[70013]: _type = "ResourcePool" [ 1123.893540] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1123.894072] env[70013]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-23dacf38-d0b5-4c40-a4b3-9de86b72a44f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.916552] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lease: (returnval){ [ 1123.916552] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52580636-4ab0-3842-fd65-f50a0b59d804" [ 1123.916552] env[70013]: _type = "HttpNfcLease" [ 1123.916552] env[70013]: } obtained for vApp import into resource pool (val){ [ 1123.916552] env[70013]: value = "resgroup-9" [ 1123.916552] env[70013]: _type = "ResourcePool" [ 1123.916552] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1123.916838] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the lease: (returnval){ [ 1123.916838] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52580636-4ab0-3842-fd65-f50a0b59d804" [ 1123.916838] env[70013]: _type = "HttpNfcLease" [ 1123.916838] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1123.928389] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1123.928389] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52580636-4ab0-3842-fd65-f50a0b59d804" [ 1123.928389] env[70013]: _type = "HttpNfcLease" [ 1123.928389] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1123.937542] env[70013]: DEBUG oslo_concurrency.lockutils [req-25b95459-56e6-4af4-8150-1face92b6276 req-5d15ceb6-17f2-46cf-9d77-99918eae52fb service nova] Releasing lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.009950] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.121974] env[70013]: DEBUG nova.network.neutron [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Successfully created port: 19122bfb-7d13-4768-b8b7-fc77264b0743 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1124.175882] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80dfe787-1ec2-4950-8dbf-2d41b8092147 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.186019] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9599b15b-b609-4a65-be4a-28466f4bc33b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.234800] env[70013]: DEBUG nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1124.242324] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7db64f-a0a4-4981-ab43-99ba9923c69a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.250614] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d46a1a-175a-42bc-89db-26f9da7bf50f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.267023] env[70013]: DEBUG nova.compute.provider_tree [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.429480] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1124.429480] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52580636-4ab0-3842-fd65-f50a0b59d804" [ 1124.429480] env[70013]: _type = "HttpNfcLease" [ 1124.429480] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1124.559954] env[70013]: DEBUG nova.network.neutron [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updated VIF entry in instance network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1124.559954] env[70013]: DEBUG nova.network.neutron [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.660591] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.660835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.770506] env[70013]: DEBUG nova.scheduler.client.report [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1124.930918] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1124.930918] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52580636-4ab0-3842-fd65-f50a0b59d804" [ 1124.930918] env[70013]: _type = "HttpNfcLease" [ 1124.930918] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1124.931387] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1124.931387] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52580636-4ab0-3842-fd65-f50a0b59d804" [ 1124.931387] env[70013]: _type = "HttpNfcLease" [ 1124.931387] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1124.932221] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5803ab-f605-4ea3-b551-f2510919b7f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.942019] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab9676-6cfa-54e5-a4c5-f9a22d5f85f3/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1124.942259] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating HTTP connection to write to file with size = 31662592 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab9676-6cfa-54e5-a4c5-f9a22d5f85f3/disk-0.vmdk. {{(pid=70013) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1125.003762] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-aa4dcf81-de3e-4163-a932-269995c3d51d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.062713] env[70013]: DEBUG oslo_concurrency.lockutils [req-5e73e59f-6588-4a6f-b7ff-2b519ec38f6f req-72781382-931c-4f28-95ef-16b9c17651ec service nova] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.165610] env[70013]: DEBUG nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1125.168729] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.168987] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.169293] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.169429] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.169600] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.172964] env[70013]: INFO nova.compute.manager [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Terminating instance [ 1125.249949] env[70013]: DEBUG nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1125.278501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.283206] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.115s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.283206] env[70013]: DEBUG nova.objects.instance [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'resources' on Instance uuid bf138a0d-ea1f-4c93-9c7f-e0e3309597b2 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.290520] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1125.290857] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.291120] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1125.291564] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.291713] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1125.292059] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1125.292210] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1125.292467] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1125.292644] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1125.292861] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1125.293115] env[70013]: DEBUG nova.virt.hardware [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1125.294076] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a10b95e-78a8-40b8-849e-96df48ae8902 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.305434] env[70013]: INFO nova.scheduler.client.report [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Deleted allocations for instance 456c4926-2ce4-4ba8-a8a6-ca50650e19f6 [ 1125.308974] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdec87c-7d72-4703-90c0-23c3d420d05c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.683533] env[70013]: DEBUG nova.compute.manager [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1125.683927] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.685528] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b782d4d-d55d-4f06-8ff7-c8cdd35cf462 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.696103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.702978] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.706664] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5407058-d836-46d5-98c9-33b5f9d731c5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.716009] env[70013]: DEBUG oslo_vmware.api [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1125.716009] env[70013]: value = "task-4231482" [ 1125.716009] env[70013]: _type = "Task" [ 1125.716009] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.724761] env[70013]: DEBUG nova.compute.manager [req-6a4562d5-3611-4a79-99ce-a50c3ca8f409 req-0f8b2c36-7dcc-4397-be29-0574f92b4c37 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Received event network-vif-plugged-19122bfb-7d13-4768-b8b7-fc77264b0743 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1125.725018] env[70013]: DEBUG oslo_concurrency.lockutils [req-6a4562d5-3611-4a79-99ce-a50c3ca8f409 req-0f8b2c36-7dcc-4397-be29-0574f92b4c37 service nova] Acquiring lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.725270] env[70013]: DEBUG oslo_concurrency.lockutils [req-6a4562d5-3611-4a79-99ce-a50c3ca8f409 req-0f8b2c36-7dcc-4397-be29-0574f92b4c37 service nova] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.725478] env[70013]: DEBUG oslo_concurrency.lockutils [req-6a4562d5-3611-4a79-99ce-a50c3ca8f409 req-0f8b2c36-7dcc-4397-be29-0574f92b4c37 service nova] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.725669] env[70013]: DEBUG nova.compute.manager [req-6a4562d5-3611-4a79-99ce-a50c3ca8f409 req-0f8b2c36-7dcc-4397-be29-0574f92b4c37 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] No waiting events found dispatching network-vif-plugged-19122bfb-7d13-4768-b8b7-fc77264b0743 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1125.725874] env[70013]: WARNING nova.compute.manager [req-6a4562d5-3611-4a79-99ce-a50c3ca8f409 req-0f8b2c36-7dcc-4397-be29-0574f92b4c37 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Received unexpected event network-vif-plugged-19122bfb-7d13-4768-b8b7-fc77264b0743 for instance with vm_state building and task_state spawning. [ 1125.731024] env[70013]: DEBUG oslo_vmware.api [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231482, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.821516] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42cec19b-768d-4606-8c32-454c609d6d64 tempest-ServerShowV257Test-1490718355 tempest-ServerShowV257Test-1490718355-project-member] Lock "456c4926-2ce4-4ba8-a8a6-ca50650e19f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.350s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.841930] env[70013]: DEBUG nova.network.neutron [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Successfully updated port: 19122bfb-7d13-4768-b8b7-fc77264b0743 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1126.215512] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5536924a-efb5-4db5-8955-e8fcaad44d10 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.234011] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c38119-4137-448f-b42f-091ecc3efe1b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.237404] env[70013]: DEBUG oslo_vmware.api [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231482, 'name': PowerOffVM_Task, 'duration_secs': 0.255615} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.240211] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.240444] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1126.241085] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43870148-9f68-41e5-9cc1-f14d00712f59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.276122] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86f2efb-8feb-41a9-b7b3-c09c7d7ff29f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.285426] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be268072-9bce-419e-b626-6be079c5e749 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.301409] env[70013]: DEBUG nova.compute.provider_tree [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.325223] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1126.325439] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1126.325622] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Deleting the datastore file [datastore2] 8dd3e9c7-3a92-49e6-883c-cda647730f5e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1126.327440] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02efb34f-2506-4869-9169-3ee10a1f36ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.335569] env[70013]: DEBUG oslo_vmware.api [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1126.335569] env[70013]: value = "task-4231484" [ 1126.335569] env[70013]: _type = "Task" [ 1126.335569] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.345425] env[70013]: DEBUG oslo_vmware.api [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.347745] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "refresh_cache-8f4d999d-54e9-4987-81f7-27a2b6d4ed17" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.347946] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquired lock "refresh_cache-8f4d999d-54e9-4987-81f7-27a2b6d4ed17" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1126.348156] env[70013]: DEBUG nova.network.neutron [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.436646] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Completed reading data from the image iterator. {{(pid=70013) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1126.437330] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab9676-6cfa-54e5-a4c5-f9a22d5f85f3/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1126.437840] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da0c4c7-5a61-435b-a4f0-7749108b9429 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.446637] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab9676-6cfa-54e5-a4c5-f9a22d5f85f3/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1126.446905] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab9676-6cfa-54e5-a4c5-f9a22d5f85f3/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1126.447175] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e619cc4f-48e7-4211-9f69-0e9cfb687c63 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.804665] env[70013]: DEBUG nova.scheduler.client.report [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1126.855454] env[70013]: DEBUG oslo_vmware.api [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.877372] env[70013]: DEBUG oslo_vmware.rw_handles [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab9676-6cfa-54e5-a4c5-f9a22d5f85f3/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1126.877611] env[70013]: INFO nova.virt.vmwareapi.images [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Downloaded image file data ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 [ 1126.878435] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1acbaa8-6413-433e-a217-33a9d0944977 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.896971] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd0434c5-3543-4730-aa20-d3c471497a30 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.899247] env[70013]: DEBUG nova.network.neutron [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1126.965955] env[70013]: INFO nova.virt.vmwareapi.images [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] The imported VM was unregistered [ 1126.968497] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Caching image {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1126.968753] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Creating directory with path [datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1126.969075] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c47f9b7-3886-4fef-8ccc-13ecb5defbcf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.996168] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Created directory with path [datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1126.996400] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632/OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632.vmdk to [datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5.vmdk. {{(pid=70013) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1126.996674] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c6fdc1d6-bcce-4e0e-b218-ab23cc227745 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.007726] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1127.007726] env[70013]: value = "task-4231486" [ 1127.007726] env[70013]: _type = "Task" [ 1127.007726] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.016306] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.062551] env[70013]: DEBUG nova.network.neutron [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Updating instance_info_cache with network_info: [{"id": "19122bfb-7d13-4768-b8b7-fc77264b0743", "address": "fa:16:3e:08:db:3a", "network": {"id": "7acaf977-8095-4101-a3b4-82576fa430e4", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1615987869-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "77d07f4b4bd24e40883d1a381ec8e002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19122bfb-7d", "ovs_interfaceid": "19122bfb-7d13-4768-b8b7-fc77264b0743", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.310620] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.029s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1127.313959] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.323s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1127.314419] env[70013]: DEBUG nova.objects.instance [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'resources' on Instance uuid 1230ca35-1512-4464-85fc-d3b4ab05eac1 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.335590] env[70013]: INFO nova.scheduler.client.report [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted allocations for instance bf138a0d-ea1f-4c93-9c7f-e0e3309597b2 [ 1127.350949] env[70013]: DEBUG oslo_vmware.api [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231484, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.532714} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.351731] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1127.351731] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1127.352009] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1127.352368] env[70013]: INFO nova.compute.manager [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Took 1.67 seconds to destroy the instance on the hypervisor. [ 1127.352691] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1127.352970] env[70013]: DEBUG nova.compute.manager [-] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1127.353093] env[70013]: DEBUG nova.network.neutron [-] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1127.517897] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.567530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Releasing lock "refresh_cache-8f4d999d-54e9-4987-81f7-27a2b6d4ed17" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.567530] env[70013]: DEBUG nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Instance network_info: |[{"id": "19122bfb-7d13-4768-b8b7-fc77264b0743", "address": "fa:16:3e:08:db:3a", "network": {"id": "7acaf977-8095-4101-a3b4-82576fa430e4", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1615987869-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "77d07f4b4bd24e40883d1a381ec8e002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19122bfb-7d", "ovs_interfaceid": "19122bfb-7d13-4768-b8b7-fc77264b0743", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1127.567530] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:db:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19122bfb-7d13-4768-b8b7-fc77264b0743', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.575689] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Creating folder: Project (77d07f4b4bd24e40883d1a381ec8e002). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1127.576049] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44e907fc-3844-49bf-9a85-9f293c063bf8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.587930] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Created folder: Project (77d07f4b4bd24e40883d1a381ec8e002) in parent group-v836999. [ 1127.588223] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Creating folder: Instances. Parent ref: group-v837222. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1127.588617] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08770936-b55c-44e0-8160-6c5fcb39116b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.599638] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Created folder: Instances in parent group-v837222. [ 1127.599907] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1127.600203] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1127.600460] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68aa88ff-a83e-4f59-8e2e-e1ac2c440250 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.623122] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.623122] env[70013]: value = "task-4231489" [ 1127.623122] env[70013]: _type = "Task" [ 1127.623122] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.636076] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231489, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.760245] env[70013]: DEBUG nova.compute.manager [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Received event network-changed-19122bfb-7d13-4768-b8b7-fc77264b0743 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1127.760613] env[70013]: DEBUG nova.compute.manager [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Refreshing instance network info cache due to event network-changed-19122bfb-7d13-4768-b8b7-fc77264b0743. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1127.760815] env[70013]: DEBUG oslo_concurrency.lockutils [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] Acquiring lock "refresh_cache-8f4d999d-54e9-4987-81f7-27a2b6d4ed17" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.760969] env[70013]: DEBUG oslo_concurrency.lockutils [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] Acquired lock "refresh_cache-8f4d999d-54e9-4987-81f7-27a2b6d4ed17" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1127.761157] env[70013]: DEBUG nova.network.neutron [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Refreshing network info cache for port 19122bfb-7d13-4768-b8b7-fc77264b0743 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.844911] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b9e3252-3f2c-4684-93d3-f66289b12756 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "bf138a0d-ea1f-4c93-9c7f-e0e3309597b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.242s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.024696] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.119475] env[70013]: DEBUG nova.network.neutron [-] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.139611] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231489, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.177337] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfd3a2d-c6a0-4faf-896b-271199654aa4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.192557] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea7e452-fa2e-426a-90bc-da14fdd3d07c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.234158] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1647c59b-b276-499c-bfc9-dcbb7ef55078 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.244828] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b061a065-63c4-4a2e-bd7e-1a3bdfa6636f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.262267] env[70013]: DEBUG nova.compute.provider_tree [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.524637] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.528511] env[70013]: DEBUG nova.network.neutron [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Updated VIF entry in instance network info cache for port 19122bfb-7d13-4768-b8b7-fc77264b0743. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.528511] env[70013]: DEBUG nova.network.neutron [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Updating instance_info_cache with network_info: [{"id": "19122bfb-7d13-4768-b8b7-fc77264b0743", "address": "fa:16:3e:08:db:3a", "network": {"id": "7acaf977-8095-4101-a3b4-82576fa430e4", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1615987869-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "77d07f4b4bd24e40883d1a381ec8e002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19122bfb-7d", "ovs_interfaceid": "19122bfb-7d13-4768-b8b7-fc77264b0743", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.625031] env[70013]: INFO nova.compute.manager [-] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Took 1.27 seconds to deallocate network for instance. [ 1128.643486] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231489, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.767509] env[70013]: DEBUG nova.scheduler.client.report [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1129.026277] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.030205] env[70013]: DEBUG oslo_concurrency.lockutils [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] Releasing lock "refresh_cache-8f4d999d-54e9-4987-81f7-27a2b6d4ed17" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1129.030358] env[70013]: DEBUG nova.compute.manager [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Received event network-vif-deleted-1d2febcb-8ea3-4126-8de7-7bbe2f7162ec {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1129.030547] env[70013]: INFO nova.compute.manager [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Neutron deleted interface 1d2febcb-8ea3-4126-8de7-7bbe2f7162ec; detaching it from the instance and deleting it from the info cache [ 1129.030827] env[70013]: DEBUG nova.network.neutron [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.138518] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.139327] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231489, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.273270] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.276426] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.746s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.276778] env[70013]: DEBUG nova.objects.instance [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lazy-loading 'resources' on Instance uuid addf8ba9-142c-4458-8586-b06e52118d71 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.365873] env[70013]: INFO nova.scheduler.client.report [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted allocations for instance 1230ca35-1512-4464-85fc-d3b4ab05eac1 [ 1129.527513] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.533976] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f9ee29c-0f44-4298-990f-a3e112641b0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.548221] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40dc2766-500e-4a79-b4ee-8044d886b534 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.582056] env[70013]: DEBUG nova.compute.manager [req-552e7963-278c-431f-a886-0911a5502906 req-e2dee50f-2ea6-4488-84cc-f7d679d22735 service nova] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Detach interface failed, port_id=1d2febcb-8ea3-4126-8de7-7bbe2f7162ec, reason: Instance 8dd3e9c7-3a92-49e6-883c-cda647730f5e could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1129.638888] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231489, 'name': CreateVM_Task, 'duration_secs': 1.938866} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.638888] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1129.639146] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.639209] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.639554] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1129.640094] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c75358d-b295-4a41-ac1b-61a03e842670 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.648777] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1129.648777] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52eb072c-1964-2603-f766-cfca6554987d" [ 1129.648777] env[70013]: _type = "Task" [ 1129.648777] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.659914] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52eb072c-1964-2603-f766-cfca6554987d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.880913] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dffbea48-64ff-4948-9780-80fa21c26503 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "1230ca35-1512-4464-85fc-d3b4ab05eac1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.382s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.026250] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.122442] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932693b2-9c62-4263-a64e-ae2fc7541a21 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.130312] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a4939c-d85e-4cc2-9259-1287ed67d897 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.165979] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a64801-5d4a-4ab9-94b5-bcf9885b1e57 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.172845] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.173093] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.183930] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbef6169-8e19-4481-8972-807013d27160 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.188291] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52eb072c-1964-2603-f766-cfca6554987d, 'name': SearchDatastore_Task, 'duration_secs': 0.097364} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.188836] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1130.189082] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1130.189326] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.189473] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1130.189649] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1130.190576] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f68d8a12-a7c3-48cf-b952-7dd8f85d2e8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.201437] env[70013]: DEBUG nova.compute.provider_tree [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.212550] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1130.212752] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1130.213587] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be686426-b4cc-4b12-b68b-0763a61e5da5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.220342] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1130.220342] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b2a71e-fd70-7f76-0217-d619c9cac9c7" [ 1130.220342] env[70013]: _type = "Task" [ 1130.220342] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.230801] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b2a71e-fd70-7f76-0217-d619c9cac9c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.524176] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231486, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.120339} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.524455] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632/OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632.vmdk to [datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5.vmdk. [ 1130.524734] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Cleaning up location [datastore1] OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1130.524906] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_94ec5f5d-95de-4ff0-b52f-01840acd1632 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.525179] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e03b3345-b7bb-4667-94a2-302a889fb730 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.532708] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1130.532708] env[70013]: value = "task-4231490" [ 1130.532708] env[70013]: _type = "Task" [ 1130.532708] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.540675] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.676673] env[70013]: DEBUG nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1130.705927] env[70013]: DEBUG nova.scheduler.client.report [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1130.731716] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b2a71e-fd70-7f76-0217-d619c9cac9c7, 'name': SearchDatastore_Task, 'duration_secs': 0.019189} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.732613] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffebf74d-aeee-44fc-9eb6-8c57f8a05d34 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.738986] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1130.738986] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52206181-f89b-88fb-a032-d0271e947e32" [ 1130.738986] env[70013]: _type = "Task" [ 1130.738986] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.747304] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52206181-f89b-88fb-a032-d0271e947e32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.043985] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231490, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039712} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.044456] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.044456] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.044726] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5.vmdk to [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1131.044998] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a02c80c-aa1e-47a3-bcda-1ed9147b5a3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.053250] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1131.053250] env[70013]: value = "task-4231491" [ 1131.053250] env[70013]: _type = "Task" [ 1131.053250] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.062661] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231491, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.212040] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.935s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.215231] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.530s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.216762] env[70013]: INFO nova.compute.claims [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1131.227737] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.250411] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52206181-f89b-88fb-a032-d0271e947e32, 'name': SearchDatastore_Task, 'duration_secs': 0.010606} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.251347] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.251347] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 8f4d999d-54e9-4987-81f7-27a2b6d4ed17/8f4d999d-54e9-4987-81f7-27a2b6d4ed17.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1131.251347] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d0ff020-3180-46fa-b137-4ef4f58fe48a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.260171] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1131.260171] env[70013]: value = "task-4231492" [ 1131.260171] env[70013]: _type = "Task" [ 1131.260171] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.261723] env[70013]: INFO nova.scheduler.client.report [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Deleted allocations for instance addf8ba9-142c-4458-8586-b06e52118d71 [ 1131.277848] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231492, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.565270] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231491, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.775888] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231492, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.776947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f320bd8a-e482-4541-9df9-cbd86779d271 tempest-ServersTestFqdnHostnames-1659575620 tempest-ServersTestFqdnHostnames-1659575620-project-member] Lock "addf8ba9-142c-4458-8586-b06e52118d71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.897s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.064852] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231491, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.276517] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231492, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.571575] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231491, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.581996] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f713f63-80c9-4d5b-862d-0d496a3e51d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.590616] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3faade54-64bb-41dc-9985-11cca25f0e28 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.631331] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0598c6f-34c5-4fa6-8b57-ac610a6eb702 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.643125] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33f7bda-4cd6-4dca-a820-8ed81f99aacc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.666051] env[70013]: DEBUG nova.compute.provider_tree [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.714289] env[70013]: DEBUG nova.compute.manager [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Stashing vm_state: active {{(pid=70013) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1132.775468] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231492, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.066601] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231491, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.170713] env[70013]: DEBUG nova.scheduler.client.report [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1133.257325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.274932] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231492, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.568140] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231491, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.677727] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.677727] env[70013]: DEBUG nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1133.681514] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.672s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.685019] env[70013]: INFO nova.compute.claims [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1133.774923] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231492, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.386231} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.775208] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 8f4d999d-54e9-4987-81f7-27a2b6d4ed17/8f4d999d-54e9-4987-81f7-27a2b6d4ed17.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1133.775428] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1133.775683] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-95793b2c-48ed-4e1e-b000-f30bd250f926 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.782766] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1133.782766] env[70013]: value = "task-4231493" [ 1133.782766] env[70013]: _type = "Task" [ 1133.782766] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.792867] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231493, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.069166] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231491, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.54838} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.069459] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5/ef9b9d1a-0aa9-4cfc-8063-b47af8f823f5.vmdk to [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1134.070339] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604a0ad2-7807-459c-944f-4e9b0d885437 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.094252] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.094587] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2964fb23-fa8f-4df9-83ad-8ba00c1d86ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.115485] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1134.115485] env[70013]: value = "task-4231494" [ 1134.115485] env[70013]: _type = "Task" [ 1134.115485] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.124937] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231494, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.191208] env[70013]: DEBUG nova.compute.utils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1134.194867] env[70013]: DEBUG nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1134.195088] env[70013]: DEBUG nova.network.neutron [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1134.253186] env[70013]: DEBUG nova.policy [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3318e4384a1741f5ad3a7c7d15384c7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d82869f19af4fbfabc03ecc3c8e59af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1134.294504] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231493, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070008} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.294504] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.295351] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad9dcf0-7b3c-4d17-ab43-622a2c1e990d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.322851] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 8f4d999d-54e9-4987-81f7-27a2b6d4ed17/8f4d999d-54e9-4987-81f7-27a2b6d4ed17.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.323201] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cab8e218-42d6-4723-aade-358607846065 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.343611] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1134.343611] env[70013]: value = "task-4231495" [ 1134.343611] env[70013]: _type = "Task" [ 1134.343611] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.355568] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231495, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.629518] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231494, 'name': ReconfigVM_Task, 'duration_secs': 0.336308} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.629861] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f/0ea84d0d-905c-428d-8abb-2781c817f08f.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.630580] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21c5bf9d-83d2-493f-80c4-17ed98edc87e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.637384] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1134.637384] env[70013]: value = "task-4231496" [ 1134.637384] env[70013]: _type = "Task" [ 1134.637384] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.645769] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231496, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.687912] env[70013]: DEBUG nova.network.neutron [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Successfully created port: cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1134.696329] env[70013]: DEBUG nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1134.858438] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231495, 'name': ReconfigVM_Task, 'duration_secs': 0.287618} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.858438] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 8f4d999d-54e9-4987-81f7-27a2b6d4ed17/8f4d999d-54e9-4987-81f7-27a2b6d4ed17.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.858438] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd24a73e-2a9f-4933-a73f-18633f5ab5b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.864971] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1134.864971] env[70013]: value = "task-4231497" [ 1134.864971] env[70013]: _type = "Task" [ 1134.864971] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.876857] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231497, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.077314] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60346c9-0216-4e83-923c-8baa19e36f6e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.085379] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b704f5bb-b9fd-49c9-b7d7-ad0501e12162 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.119019] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2710dbd3-430f-4ec5-b807-00c13165686f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.126670] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e72ccf0-22d4-4ad6-a30b-86cfa4e02deb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.141982] env[70013]: DEBUG nova.compute.provider_tree [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.151713] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231496, 'name': Rename_Task, 'duration_secs': 0.191734} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.152614] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.152855] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc1a6d45-3290-45e5-8c79-7b04c27cf1d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.160857] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1135.160857] env[70013]: value = "task-4231498" [ 1135.160857] env[70013]: _type = "Task" [ 1135.160857] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.172359] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.375722] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231497, 'name': Rename_Task, 'duration_secs': 0.157549} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.376068] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.376354] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11ab497c-11ed-4736-897a-fe9f81545998 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.383165] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1135.383165] env[70013]: value = "task-4231499" [ 1135.383165] env[70013]: _type = "Task" [ 1135.383165] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.391747] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.649023] env[70013]: DEBUG nova.scheduler.client.report [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1135.674623] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231498, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.709914] env[70013]: DEBUG nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1135.770283] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1135.770559] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1135.770719] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1135.770900] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1135.771098] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1135.771263] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1135.771486] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1135.771673] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1135.771846] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1135.772011] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1135.772208] env[70013]: DEBUG nova.virt.hardware [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1135.774910] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcd6057-3b3a-40e7-be80-467a0da6f610 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.783109] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e874a38c-d077-497f-ac42-920a0352505c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.894298] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231499, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.157974] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.157974] env[70013]: DEBUG nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1136.158642] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.463s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.160660] env[70013]: INFO nova.compute.claims [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1136.172497] env[70013]: DEBUG oslo_vmware.api [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231498, 'name': PowerOnVM_Task, 'duration_secs': 0.625879} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.173597] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.367274] env[70013]: DEBUG nova.compute.manager [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1136.368222] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4acd43b2-06a8-4d38-a8c9-37c0429efa70 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.394132] env[70013]: DEBUG oslo_vmware.api [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231499, 'name': PowerOnVM_Task, 'duration_secs': 0.567151} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.394447] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.394666] env[70013]: INFO nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Took 11.14 seconds to spawn the instance on the hypervisor. [ 1136.395041] env[70013]: DEBUG nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1136.396022] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ece239-97ba-4c84-b80c-c7b7b03d0527 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.447329] env[70013]: DEBUG nova.compute.manager [req-51641c72-c77f-44ee-8134-5ebcfba51105 req-a4ad788b-5793-4ab5-ba9d-ab96a09590ad service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Received event network-vif-plugged-cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1136.447571] env[70013]: DEBUG oslo_concurrency.lockutils [req-51641c72-c77f-44ee-8134-5ebcfba51105 req-a4ad788b-5793-4ab5-ba9d-ab96a09590ad service nova] Acquiring lock "499bb115-146c-48dc-9447-3c01c9549a73-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.447782] env[70013]: DEBUG oslo_concurrency.lockutils [req-51641c72-c77f-44ee-8134-5ebcfba51105 req-a4ad788b-5793-4ab5-ba9d-ab96a09590ad service nova] Lock "499bb115-146c-48dc-9447-3c01c9549a73-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.447954] env[70013]: DEBUG oslo_concurrency.lockutils [req-51641c72-c77f-44ee-8134-5ebcfba51105 req-a4ad788b-5793-4ab5-ba9d-ab96a09590ad service nova] Lock "499bb115-146c-48dc-9447-3c01c9549a73-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.448140] env[70013]: DEBUG nova.compute.manager [req-51641c72-c77f-44ee-8134-5ebcfba51105 req-a4ad788b-5793-4ab5-ba9d-ab96a09590ad service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] No waiting events found dispatching network-vif-plugged-cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1136.448323] env[70013]: WARNING nova.compute.manager [req-51641c72-c77f-44ee-8134-5ebcfba51105 req-a4ad788b-5793-4ab5-ba9d-ab96a09590ad service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Received unexpected event network-vif-plugged-cc3af4a9-2d59-4b89-975c-87e4a5e6d939 for instance with vm_state building and task_state spawning. [ 1136.664777] env[70013]: DEBUG nova.compute.utils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1136.668129] env[70013]: DEBUG nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1136.668342] env[70013]: DEBUG nova.network.neutron [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1136.721091] env[70013]: DEBUG nova.policy [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfa9f992483942f689f616b60a27b00d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0896b76438d9456fb1b9fa2f39d0decc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1136.814457] env[70013]: DEBUG nova.network.neutron [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Successfully updated port: cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1136.832628] env[70013]: DEBUG nova.compute.manager [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Received event network-changed-cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1136.832949] env[70013]: DEBUG nova.compute.manager [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Refreshing instance network info cache due to event network-changed-cc3af4a9-2d59-4b89-975c-87e4a5e6d939. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1136.833115] env[70013]: DEBUG oslo_concurrency.lockutils [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] Acquiring lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.833278] env[70013]: DEBUG oslo_concurrency.lockutils [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] Acquired lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.833459] env[70013]: DEBUG nova.network.neutron [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Refreshing network info cache for port cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1136.885990] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f807ff5-7210-4577-90e7-61c263f59cc2 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 31.370s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.913922] env[70013]: INFO nova.compute.manager [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Took 29.57 seconds to build instance. [ 1137.145047] env[70013]: DEBUG nova.network.neutron [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Successfully created port: 3eb52fcc-22d7-430e-9fd7-54b925e05cec {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1137.174521] env[70013]: DEBUG nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1137.316390] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.390997] env[70013]: DEBUG nova.network.neutron [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1137.417354] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89d2df96-1fff-4baa-9210-1e1c045d6365 tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.086s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1137.531954] env[70013]: DEBUG nova.network.neutron [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.566829] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d554ca-8b7d-44d7-aa9a-7841a9329f58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.576351] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc90b96-6b01-4bb5-a234-6791e6b84b07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.619072] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbbe54c-cd3f-4381-a8e9-3b81d869fd0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.629317] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20deea4d-4311-4cfa-80d3-5ed30fdbef98 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.649112] env[70013]: DEBUG nova.compute.provider_tree [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.034921] env[70013]: DEBUG oslo_concurrency.lockutils [req-7516dbed-dfb7-4fc8-a25f-8fc47face1c8 req-3aa54938-c304-4fd7-8c7e-bc402347705f service nova] Releasing lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1138.035377] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquired lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1138.035628] env[70013]: DEBUG nova.network.neutron [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.152240] env[70013]: DEBUG nova.scheduler.client.report [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1138.189803] env[70013]: DEBUG nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1138.240411] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1138.240660] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1138.240820] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1138.241035] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1138.241169] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1138.241315] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1138.241530] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1138.241690] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1138.241856] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1138.242033] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1138.242219] env[70013]: DEBUG nova.virt.hardware [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1138.243481] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7fa4bf-8c1d-4f16-88f4-f5f1a87e64fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.252446] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e848b679-bfc2-4bbe-9c3f-0f057705d489 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.574198] env[70013]: DEBUG nova.network.neutron [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1138.660013] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.660597] env[70013]: DEBUG nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1138.668748] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.529s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.669035] env[70013]: DEBUG nova.objects.instance [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lazy-loading 'resources' on Instance uuid 8dd3e9c7-3a92-49e6-883c-cda647730f5e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.820468] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.820770] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.820987] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.821196] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.821403] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.823972] env[70013]: INFO nova.compute.manager [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Terminating instance [ 1138.833861] env[70013]: DEBUG nova.network.neutron [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Updating instance_info_cache with network_info: [{"id": "cc3af4a9-2d59-4b89-975c-87e4a5e6d939", "address": "fa:16:3e:4b:e4:7f", "network": {"id": "8f216d33-a334-43c6-b03f-fc45ad272a21", "bridge": "br-int", "label": "tempest-ServersTestJSON-1533284893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d82869f19af4fbfabc03ecc3c8e59af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc3af4a9-2d", "ovs_interfaceid": "cc3af4a9-2d59-4b89-975c-87e4a5e6d939", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.857348] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb93bcf-982a-4061-85e0-9c086d160cfe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.865388] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5395cde1-62e2-4d5b-ae35-74f24df3a6af tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Suspending the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1138.865653] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6956f96f-f88c-40c4-9355-3af4cec2210c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.873186] env[70013]: DEBUG oslo_vmware.api [None req-5395cde1-62e2-4d5b-ae35-74f24df3a6af tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1138.873186] env[70013]: value = "task-4231500" [ 1138.873186] env[70013]: _type = "Task" [ 1138.873186] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.882333] env[70013]: DEBUG oslo_vmware.api [None req-5395cde1-62e2-4d5b-ae35-74f24df3a6af tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231500, 'name': SuspendVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.937620] env[70013]: DEBUG nova.compute.manager [req-819d4f36-e212-4673-a0f5-9817f882f903 req-b085d875-cdec-4695-806a-b65f1b10696e service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Received event network-vif-plugged-3eb52fcc-22d7-430e-9fd7-54b925e05cec {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1138.937838] env[70013]: DEBUG oslo_concurrency.lockutils [req-819d4f36-e212-4673-a0f5-9817f882f903 req-b085d875-cdec-4695-806a-b65f1b10696e service nova] Acquiring lock "e11550bf-3284-4ae6-9aa9-34923da589ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.938076] env[70013]: DEBUG oslo_concurrency.lockutils [req-819d4f36-e212-4673-a0f5-9817f882f903 req-b085d875-cdec-4695-806a-b65f1b10696e service nova] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.938318] env[70013]: DEBUG oslo_concurrency.lockutils [req-819d4f36-e212-4673-a0f5-9817f882f903 req-b085d875-cdec-4695-806a-b65f1b10696e service nova] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.938425] env[70013]: DEBUG nova.compute.manager [req-819d4f36-e212-4673-a0f5-9817f882f903 req-b085d875-cdec-4695-806a-b65f1b10696e service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] No waiting events found dispatching network-vif-plugged-3eb52fcc-22d7-430e-9fd7-54b925e05cec {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1138.938549] env[70013]: WARNING nova.compute.manager [req-819d4f36-e212-4673-a0f5-9817f882f903 req-b085d875-cdec-4695-806a-b65f1b10696e service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Received unexpected event network-vif-plugged-3eb52fcc-22d7-430e-9fd7-54b925e05cec for instance with vm_state building and task_state spawning. [ 1139.028344] env[70013]: DEBUG nova.network.neutron [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Successfully updated port: 3eb52fcc-22d7-430e-9fd7-54b925e05cec {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1139.172633] env[70013]: DEBUG nova.compute.utils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1139.174316] env[70013]: DEBUG nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1139.174546] env[70013]: DEBUG nova.network.neutron [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1139.225102] env[70013]: DEBUG nova.policy [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1139.328772] env[70013]: DEBUG nova.compute.manager [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1139.329027] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1139.330238] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bde0fed-f6f4-4804-af48-0ea91dddff90 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.339269] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Releasing lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1139.339269] env[70013]: DEBUG nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Instance network_info: |[{"id": "cc3af4a9-2d59-4b89-975c-87e4a5e6d939", "address": "fa:16:3e:4b:e4:7f", "network": {"id": "8f216d33-a334-43c6-b03f-fc45ad272a21", "bridge": "br-int", "label": "tempest-ServersTestJSON-1533284893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d82869f19af4fbfabc03ecc3c8e59af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc3af4a9-2d", "ovs_interfaceid": "cc3af4a9-2d59-4b89-975c-87e4a5e6d939", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1139.341977] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:e4:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc3af4a9-2d59-4b89-975c-87e4a5e6d939', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.349585] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Creating folder: Project (6d82869f19af4fbfabc03ecc3c8e59af). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1139.349932] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1139.350196] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae2adadd-2c01-4fea-a495-f42fd7b9494a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.352198] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e89166d-682b-44e1-9308-967811b7512d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.362585] env[70013]: DEBUG oslo_vmware.api [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1139.362585] env[70013]: value = "task-4231501" [ 1139.362585] env[70013]: _type = "Task" [ 1139.362585] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.370604] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Created folder: Project (6d82869f19af4fbfabc03ecc3c8e59af) in parent group-v836999. [ 1139.370824] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Creating folder: Instances. Parent ref: group-v837225. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1139.371754] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bf90bfa-3fd5-43fd-8082-dc7799edd219 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.381105] env[70013]: DEBUG oslo_vmware.api [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.386400] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Created folder: Instances in parent group-v837225. [ 1139.386661] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1139.389725] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1139.390028] env[70013]: DEBUG oslo_vmware.api [None req-5395cde1-62e2-4d5b-ae35-74f24df3a6af tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231500, 'name': SuspendVM_Task} progress is 58%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.390258] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18be4af2-e152-4568-bc19-4c7ffc7eff20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.414270] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1139.414270] env[70013]: value = "task-4231504" [ 1139.414270] env[70013]: _type = "Task" [ 1139.414270] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.425224] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231504, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.535095] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-e11550bf-3284-4ae6-9aa9-34923da589ca" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.535095] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-e11550bf-3284-4ae6-9aa9-34923da589ca" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1139.535095] env[70013]: DEBUG nova.network.neutron [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1139.575061] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af7cdd2-c3f0-4fc2-baac-aebc10d52254 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.585454] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c2f2a6-aea9-4fb9-80d3-afb0d89394e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.590116] env[70013]: DEBUG nova.network.neutron [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Successfully created port: dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1139.620364] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1efebf-ad4e-4635-a2e1-932c17cfbfdf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.629114] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ce8ec7-a0a6-4dc4-aaff-5cfd7c514f90 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.646242] env[70013]: DEBUG nova.compute.provider_tree [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.680550] env[70013]: DEBUG nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1139.873719] env[70013]: DEBUG oslo_vmware.api [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231501, 'name': PowerOffVM_Task, 'duration_secs': 0.296107} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.874137] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1139.874319] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1139.877404] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-edbc43cc-c5e0-4b11-ba3a-a8e799285594 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.885747] env[70013]: DEBUG oslo_vmware.api [None req-5395cde1-62e2-4d5b-ae35-74f24df3a6af tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231500, 'name': SuspendVM_Task, 'duration_secs': 0.71795} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.886256] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5395cde1-62e2-4d5b-ae35-74f24df3a6af tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Suspended the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1139.886546] env[70013]: DEBUG nova.compute.manager [None req-5395cde1-62e2-4d5b-ae35-74f24df3a6af tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1139.887574] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec22f8b-3b4f-4908-bef6-a8bb3e0d53fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.924420] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231504, 'name': CreateVM_Task, 'duration_secs': 0.466423} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.924652] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1139.925414] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.925586] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1139.925910] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1139.926189] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c52b66fb-fd17-4f28-a023-e455647f04ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.931965] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1139.931965] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b4b4c7-7741-1fa1-81e6-04e784220ce1" [ 1139.931965] env[70013]: _type = "Task" [ 1139.931965] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.941311] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b4b4c7-7741-1fa1-81e6-04e784220ce1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.071720] env[70013]: DEBUG nova.network.neutron [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1140.149316] env[70013]: DEBUG nova.scheduler.client.report [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1140.200873] env[70013]: DEBUG nova.network.neutron [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Updating instance_info_cache with network_info: [{"id": "3eb52fcc-22d7-430e-9fd7-54b925e05cec", "address": "fa:16:3e:61:02:f8", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb52fcc-22", "ovs_interfaceid": "3eb52fcc-22d7-430e-9fd7-54b925e05cec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.352425] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1140.352425] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1140.352607] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Deleting the datastore file [datastore1] 8f4d999d-54e9-4987-81f7-27a2b6d4ed17 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1140.352808] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50fe4d14-3cf1-448a-a660-e7c10a76501a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.360272] env[70013]: DEBUG oslo_vmware.api [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for the task: (returnval){ [ 1140.360272] env[70013]: value = "task-4231506" [ 1140.360272] env[70013]: _type = "Task" [ 1140.360272] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.369410] env[70013]: DEBUG oslo_vmware.api [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231506, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.443529] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b4b4c7-7741-1fa1-81e6-04e784220ce1, 'name': SearchDatastore_Task, 'duration_secs': 0.010572} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.443894] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1140.444150] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1140.444397] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.444546] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1140.444779] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1140.445087] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4c142f6-24cb-4eb1-b745-bb6e40e715e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.454893] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1140.454893] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1140.455675] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f839de6-5e2f-4de4-94e7-1ee1bee3542c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.461765] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1140.461765] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52114da4-39f1-907b-53db-f0d8e2c5a8d4" [ 1140.461765] env[70013]: _type = "Task" [ 1140.461765] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.470325] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52114da4-39f1-907b-53db-f0d8e2c5a8d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.655081] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.658037] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.429s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.661407] env[70013]: INFO nova.compute.claims [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.691018] env[70013]: DEBUG nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1140.704595] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-e11550bf-3284-4ae6-9aa9-34923da589ca" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1140.705021] env[70013]: DEBUG nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Instance network_info: |[{"id": "3eb52fcc-22d7-430e-9fd7-54b925e05cec", "address": "fa:16:3e:61:02:f8", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb52fcc-22", "ovs_interfaceid": "3eb52fcc-22d7-430e-9fd7-54b925e05cec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1140.705900] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:02:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac7039c0-3374-4c08-87fc-af2449b48b02', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3eb52fcc-22d7-430e-9fd7-54b925e05cec', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1140.714253] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1140.715446] env[70013]: INFO nova.scheduler.client.report [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Deleted allocations for instance 8dd3e9c7-3a92-49e6-883c-cda647730f5e [ 1140.720306] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1140.720787] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-592507e2-76cd-458a-9ce8-8dd6dec44283 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.745678] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1140.746036] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1140.746036] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1140.746255] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1140.746448] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1140.746521] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1140.746741] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1140.746875] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1140.747037] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1140.747265] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1140.747469] env[70013]: DEBUG nova.virt.hardware [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1140.749018] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32976884-9a17-47b6-a7a1-a31437077623 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.754041] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1140.754041] env[70013]: value = "task-4231507" [ 1140.754041] env[70013]: _type = "Task" [ 1140.754041] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.761215] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577ee414-1cb4-426f-8138-e903ac0b55bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.768922] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231507, 'name': CreateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.870626] env[70013]: DEBUG oslo_vmware.api [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Task: {'id': task-4231506, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398778} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.870895] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1140.871137] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1140.871334] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1140.871518] env[70013]: INFO nova.compute.manager [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Took 1.54 seconds to destroy the instance on the hypervisor. [ 1140.871876] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1140.872119] env[70013]: DEBUG nova.compute.manager [-] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1140.872220] env[70013]: DEBUG nova.network.neutron [-] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1140.977898] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52114da4-39f1-907b-53db-f0d8e2c5a8d4, 'name': SearchDatastore_Task, 'duration_secs': 0.010748} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.977898] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a61c862d-61f8-4840-a960-0e562ab34a2d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.981764] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1140.981764] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529337f8-678f-075a-2c4b-e0cef33eea5d" [ 1140.981764] env[70013]: _type = "Task" [ 1140.981764] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.991522] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529337f8-678f-075a-2c4b-e0cef33eea5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.101761] env[70013]: DEBUG nova.compute.manager [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Received event network-changed-3eb52fcc-22d7-430e-9fd7-54b925e05cec {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1141.101761] env[70013]: DEBUG nova.compute.manager [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Refreshing instance network info cache due to event network-changed-3eb52fcc-22d7-430e-9fd7-54b925e05cec. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1141.101761] env[70013]: DEBUG oslo_concurrency.lockutils [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] Acquiring lock "refresh_cache-e11550bf-3284-4ae6-9aa9-34923da589ca" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.101761] env[70013]: DEBUG oslo_concurrency.lockutils [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] Acquired lock "refresh_cache-e11550bf-3284-4ae6-9aa9-34923da589ca" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.101761] env[70013]: DEBUG nova.network.neutron [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Refreshing network info cache for port 3eb52fcc-22d7-430e-9fd7-54b925e05cec {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1141.181819] env[70013]: DEBUG nova.compute.manager [req-d25e67f5-092b-4f12-9413-5f82751b6cb2 req-489a7be9-6e8c-4e1c-9bf9-7b7901019e5a service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-vif-plugged-dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1141.182209] env[70013]: DEBUG oslo_concurrency.lockutils [req-d25e67f5-092b-4f12-9413-5f82751b6cb2 req-489a7be9-6e8c-4e1c-9bf9-7b7901019e5a service nova] Acquiring lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.183921] env[70013]: DEBUG oslo_concurrency.lockutils [req-d25e67f5-092b-4f12-9413-5f82751b6cb2 req-489a7be9-6e8c-4e1c-9bf9-7b7901019e5a service nova] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.183921] env[70013]: DEBUG oslo_concurrency.lockutils [req-d25e67f5-092b-4f12-9413-5f82751b6cb2 req-489a7be9-6e8c-4e1c-9bf9-7b7901019e5a service nova] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.183921] env[70013]: DEBUG nova.compute.manager [req-d25e67f5-092b-4f12-9413-5f82751b6cb2 req-489a7be9-6e8c-4e1c-9bf9-7b7901019e5a service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] No waiting events found dispatching network-vif-plugged-dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1141.183921] env[70013]: WARNING nova.compute.manager [req-d25e67f5-092b-4f12-9413-5f82751b6cb2 req-489a7be9-6e8c-4e1c-9bf9-7b7901019e5a service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received unexpected event network-vif-plugged-dcb5c6dd-7799-4712-91fd-86a53ac56170 for instance with vm_state building and task_state spawning. [ 1141.225736] env[70013]: DEBUG oslo_concurrency.lockutils [None req-58cbf2ef-979c-4dc8-8dbc-54614d05c889 tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "8dd3e9c7-3a92-49e6-883c-cda647730f5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.056s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.269998] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231507, 'name': CreateVM_Task, 'duration_secs': 0.46414} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.270249] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1141.270953] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.271132] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.271485] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1141.271751] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e8e3027-d8b0-48c9-aadf-ab6e33f16b58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.277353] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1141.277353] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529e86c2-df6a-a82e-b107-eb275a21cbc6" [ 1141.277353] env[70013]: _type = "Task" [ 1141.277353] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.286212] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529e86c2-df6a-a82e-b107-eb275a21cbc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.492030] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529337f8-678f-075a-2c4b-e0cef33eea5d, 'name': SearchDatastore_Task, 'duration_secs': 0.011435} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.492367] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1141.492627] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 499bb115-146c-48dc-9447-3c01c9549a73/499bb115-146c-48dc-9447-3c01c9549a73.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1141.492905] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39bc5a24-167d-43c1-ba45-189727874693 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.502065] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1141.502065] env[70013]: value = "task-4231508" [ 1141.502065] env[70013]: _type = "Task" [ 1141.502065] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.511917] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.537746] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.537991] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.817616] env[70013]: DEBUG nova.network.neutron [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Successfully updated port: dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "c57e876a-8696-4ce5-b533-7580f484414e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "c57e876a-8696-4ce5-b533-7580f484414e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "c57e876a-8696-4ce5-b533-7580f484414e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "c57e876a-8696-4ce5-b533-7580f484414e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "c57e876a-8696-4ce5-b533-7580f484414e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.817616] env[70013]: INFO nova.compute.manager [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Terminating instance [ 1141.817616] env[70013]: INFO nova.compute.manager [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Resuming [ 1141.817616] env[70013]: DEBUG nova.objects.instance [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lazy-loading 'flavor' on Instance uuid 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.817616] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529e86c2-df6a-a82e-b107-eb275a21cbc6, 'name': SearchDatastore_Task, 'duration_secs': 0.01018} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1141.817616] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.817616] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.820150] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1141.820150] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-694f6d0a-c987-4eac-a01d-c7b483124e86 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.820150] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1141.820150] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1141.823082] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c77a585-a06b-4549-9a95-d0c3099c3796 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.834734] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1141.834734] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52316ac9-62b9-693c-4d69-904965dc4c84" [ 1141.834734] env[70013]: _type = "Task" [ 1141.834734] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.853876] env[70013]: DEBUG nova.network.neutron [-] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.855386] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52316ac9-62b9-693c-4d69-904965dc4c84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.014487] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488539} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.014763] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 499bb115-146c-48dc-9447-3c01c9549a73/499bb115-146c-48dc-9447-3c01c9549a73.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1142.015226] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1142.015226] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a22c03ae-351e-40c6-a94d-d354b62e167c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.027760] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1142.027760] env[70013]: value = "task-4231509" [ 1142.027760] env[70013]: _type = "Task" [ 1142.027760] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.040147] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.055034] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.055034] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.055034] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.055034] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.055034] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.055034] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.055034] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1142.055034] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.091887] env[70013]: DEBUG nova.network.neutron [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Updated VIF entry in instance network info cache for port 3eb52fcc-22d7-430e-9fd7-54b925e05cec. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1142.091887] env[70013]: DEBUG nova.network.neutron [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Updating instance_info_cache with network_info: [{"id": "3eb52fcc-22d7-430e-9fd7-54b925e05cec", "address": "fa:16:3e:61:02:f8", "network": {"id": "7e24c2f4-0563-49cb-9477-2d8b5968a69e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1208173494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0896b76438d9456fb1b9fa2f39d0decc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb52fcc-22", "ovs_interfaceid": "3eb52fcc-22d7-430e-9fd7-54b925e05cec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.153269] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f92a9c8-6be5-4cb3-a897-86ac00c26591 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.161294] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f52c14-6f7d-4108-bcc4-b9a40ead18e5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.193144] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f885f89d-c0e7-43bc-8bd5-0be735784179 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.206053] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60541999-00ee-4c88-a194-906745a0d2d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.220320] env[70013]: DEBUG nova.compute.provider_tree [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.242808] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.243147] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.243147] env[70013]: DEBUG nova.network.neutron [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1142.260746] env[70013]: DEBUG nova.compute.manager [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1142.261080] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1142.263549] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c8004a-d490-471d-a664-f53afbb08257 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.272569] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1142.272859] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe924f5a-258b-4091-aae2-238993ec161c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.281399] env[70013]: DEBUG oslo_vmware.api [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1142.281399] env[70013]: value = "task-4231510" [ 1142.281399] env[70013]: _type = "Task" [ 1142.281399] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.291876] env[70013]: DEBUG oslo_vmware.api [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.347010] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52316ac9-62b9-693c-4d69-904965dc4c84, 'name': SearchDatastore_Task, 'duration_secs': 0.058364} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.347924] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-192ee6af-909a-400c-9a09-d2792a45d189 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.355089] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1142.355089] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd64fe-d3aa-07eb-c4d3-96d44a4e4d81" [ 1142.355089] env[70013]: _type = "Task" [ 1142.355089] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.359338] env[70013]: INFO nova.compute.manager [-] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Took 1.49 seconds to deallocate network for instance. [ 1142.368283] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd64fe-d3aa-07eb-c4d3-96d44a4e4d81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.539483] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06815} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.539792] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1142.540687] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32dc026-5c56-462f-94e4-e19d24c51337 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.556984] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.566017] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 499bb115-146c-48dc-9447-3c01c9549a73/499bb115-146c-48dc-9447-3c01c9549a73.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1142.566416] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e1610b7-a171-4af4-bd6d-cd6304f6fa9d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.587590] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1142.587590] env[70013]: value = "task-4231511" [ 1142.587590] env[70013]: _type = "Task" [ 1142.587590] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.591477] env[70013]: DEBUG oslo_concurrency.lockutils [req-26472480-6be2-45f0-9744-3de8e409caf1 req-2cab43f0-f0ea-439a-9420-22b9c93ab8c4 service nova] Releasing lock "refresh_cache-e11550bf-3284-4ae6-9aa9-34923da589ca" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.597495] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231511, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.723526] env[70013]: DEBUG nova.scheduler.client.report [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.776506] env[70013]: DEBUG nova.network.neutron [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1142.791674] env[70013]: DEBUG oslo_vmware.api [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231510, 'name': PowerOffVM_Task, 'duration_secs': 0.220782} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.791938] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1142.792124] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1142.792384] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d40026c-5f14-4653-becd-c05c33fa84ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.860976] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1142.861235] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1142.861411] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Deleting the datastore file [datastore2] c57e876a-8696-4ce5-b533-7580f484414e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1142.864672] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6c122a9-8340-4cca-9c1c-5b8a7f933c50 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.872791] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.873083] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bd64fe-d3aa-07eb-c4d3-96d44a4e4d81, 'name': SearchDatastore_Task, 'duration_secs': 0.012987} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.874369] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.874631] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e11550bf-3284-4ae6-9aa9-34923da589ca/e11550bf-3284-4ae6-9aa9-34923da589ca.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1142.874967] env[70013]: DEBUG oslo_vmware.api [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for the task: (returnval){ [ 1142.874967] env[70013]: value = "task-4231513" [ 1142.874967] env[70013]: _type = "Task" [ 1142.874967] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.875168] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2dd70a5-55d6-4895-a7b5-e4dd54163258 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.885534] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1142.885534] env[70013]: value = "task-4231514" [ 1142.885534] env[70013]: _type = "Task" [ 1142.885534] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.892860] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.928622] env[70013]: DEBUG nova.network.neutron [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.101387] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231511, 'name': ReconfigVM_Task, 'duration_secs': 0.257938} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.101739] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 499bb115-146c-48dc-9447-3c01c9549a73/499bb115-146c-48dc-9447-3c01c9549a73.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1143.102416] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0a09bc7-c99b-4883-bf27-61736f096046 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.112095] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1143.112095] env[70013]: value = "task-4231515" [ 1143.112095] env[70013]: _type = "Task" [ 1143.112095] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.125262] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231515, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.132421] env[70013]: DEBUG nova.compute.manager [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Received event network-vif-deleted-19122bfb-7d13-4768-b8b7-fc77264b0743 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1143.132667] env[70013]: DEBUG nova.compute.manager [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1143.132835] env[70013]: DEBUG nova.compute.manager [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing instance network info cache due to event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1143.133468] env[70013]: DEBUG oslo_concurrency.lockutils [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.229647] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.230470] env[70013]: DEBUG nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1143.234254] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 9.977s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1143.270639] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.271142] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquired lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.271142] env[70013]: DEBUG nova.network.neutron [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1143.388021] env[70013]: DEBUG oslo_vmware.api [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Task: {'id': task-4231513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.485819} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.391403] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.391613] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1143.391798] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1143.392041] env[70013]: INFO nova.compute.manager [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1143.392314] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1143.392895] env[70013]: DEBUG nova.compute.manager [-] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1143.393056] env[70013]: DEBUG nova.network.neutron [-] [instance: c57e876a-8696-4ce5-b533-7580f484414e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1143.400257] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483712} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.400544] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e11550bf-3284-4ae6-9aa9-34923da589ca/e11550bf-3284-4ae6-9aa9-34923da589ca.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1143.400783] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1143.401045] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0dab0ae8-5e70-468c-9c91-3aff1c4af22d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.409371] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1143.409371] env[70013]: value = "task-4231516" [ 1143.409371] env[70013]: _type = "Task" [ 1143.409371] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.418774] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231516, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.431667] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.432373] env[70013]: DEBUG nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Instance network_info: |[{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1143.432629] env[70013]: DEBUG oslo_concurrency.lockutils [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.432916] env[70013]: DEBUG nova.network.neutron [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1143.434817] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:9a:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dcb5c6dd-7799-4712-91fd-86a53ac56170', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1143.443405] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1143.444435] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1143.444673] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41df2780-8358-4356-a176-0dda21029f73 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.467803] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1143.467803] env[70013]: value = "task-4231517" [ 1143.467803] env[70013]: _type = "Task" [ 1143.467803] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.478103] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231517, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.623701] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231515, 'name': Rename_Task, 'duration_secs': 0.321549} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.623995] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1143.624246] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd5585a6-50a2-4985-b86e-13720f4b47d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.632784] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1143.632784] env[70013]: value = "task-4231518" [ 1143.632784] env[70013]: _type = "Task" [ 1143.632784] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.642030] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231518, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.738371] env[70013]: DEBUG nova.compute.utils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1143.742299] env[70013]: INFO nova.compute.claims [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1143.746252] env[70013]: DEBUG nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1143.746560] env[70013]: DEBUG nova.network.neutron [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1143.857193] env[70013]: DEBUG nova.policy [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3bcb0d0bd1e44eb960d676070189f27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3d42e73dc0a4a06a2022d8b54e13e12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1143.920830] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231516, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071914} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.921139] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1143.921930] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9279b8a1-3251-4eca-b566-e7b265e5365b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.948325] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] e11550bf-3284-4ae6-9aa9-34923da589ca/e11550bf-3284-4ae6-9aa9-34923da589ca.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1143.951640] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d69c183b-8dc8-4ebd-b0bf-d32375cc6ddd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.974790] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1143.974790] env[70013]: value = "task-4231519" [ 1143.974790] env[70013]: _type = "Task" [ 1143.974790] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.980907] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231517, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.990090] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231519, 'name': ReconfigVM_Task} progress is 10%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.083277] env[70013]: DEBUG nova.network.neutron [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [{"id": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "address": "fa:16:3e:8f:be:a6", "network": {"id": "68e733e9-702a-4958-baff-d27df606d709", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1103313558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68ad346a8e3a465da70d3a7de825ac6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8aacd8d-2f", "ovs_interfaceid": "d8aacd8d-2fc2-477b-9d82-e101e20b0132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.145073] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231518, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.175279] env[70013]: DEBUG nova.network.neutron [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updated VIF entry in instance network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1144.175649] env[70013]: DEBUG nova.network.neutron [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.251875] env[70013]: INFO nova.compute.resource_tracker [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating resource usage from migration 8f1251c5-4dcb-4b6f-afa5-81f973aaec05 [ 1144.254709] env[70013]: DEBUG nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1144.266187] env[70013]: DEBUG nova.network.neutron [-] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.269047] env[70013]: DEBUG nova.network.neutron [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Successfully created port: 43a074aa-e66b-4258-8c74-fae23dca9c71 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1144.482175] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231517, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.493489] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.586142] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Releasing lock "refresh_cache-0ea84d0d-905c-428d-8abb-2781c817f08f" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.587448] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6250ff-dd81-4534-957c-b2bd757f7e15 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.594552] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Resuming the VM {{(pid=70013) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1144.594897] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab711db7-8886-4cbc-90d1-17cffea0f42b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.604218] env[70013]: DEBUG oslo_vmware.api [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1144.604218] env[70013]: value = "task-4231520" [ 1144.604218] env[70013]: _type = "Task" [ 1144.604218] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.615252] env[70013]: DEBUG oslo_vmware.api [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231520, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.624299] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44c6434-c74d-4c3e-811c-1b9457de8d93 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.639496] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e09512-1a82-4a40-99c5-b2df04360cd4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.648893] env[70013]: DEBUG oslo_vmware.api [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231518, 'name': PowerOnVM_Task, 'duration_secs': 0.691247} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.677318] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1144.677667] env[70013]: INFO nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Took 8.97 seconds to spawn the instance on the hypervisor. [ 1144.677879] env[70013]: DEBUG nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1144.678768] env[70013]: DEBUG oslo_concurrency.lockutils [req-f95dce83-fa1f-49be-bb9c-67ca9ea7b80e req-8f024a18-e510-4d4b-8937-93dd144b5e96 service nova] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.679648] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d68700-5eed-4c11-b64a-639e42353b40 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.682911] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db50ae2-e2a4-4ce8-8d63-b0bd638b07cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.691877] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00aaf109-3e5e-49f9-bb0c-1f1480461040 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.708867] env[70013]: DEBUG nova.compute.provider_tree [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.771689] env[70013]: INFO nova.compute.manager [-] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Took 1.38 seconds to deallocate network for instance. [ 1144.982320] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231517, 'name': CreateVM_Task, 'duration_secs': 1.413256} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.985352] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1144.986085] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.986258] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.986570] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1144.987176] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f35204e-eb8f-4f85-a0da-0dd06bcca827 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.992888] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.993515] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1144.993515] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523920b4-3729-7490-c46e-631329266243" [ 1144.993515] env[70013]: _type = "Task" [ 1144.993515] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.001243] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523920b4-3729-7490-c46e-631329266243, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.115264] env[70013]: DEBUG oslo_vmware.api [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231520, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.166613] env[70013]: DEBUG nova.compute.manager [req-00ef928d-2464-4b74-aa00-616d5f03bce6 req-7f7b7f62-c9a1-4294-ba31-1eb7997ead96 service nova] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Received event network-vif-deleted-f27a5efc-f9a7-4cb7-8db5-e1e692932fe1 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1145.209231] env[70013]: INFO nova.compute.manager [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Took 21.54 seconds to build instance. [ 1145.212241] env[70013]: DEBUG nova.scheduler.client.report [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1145.267567] env[70013]: DEBUG nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1145.278863] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1145.325234] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1145.325529] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.325699] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1145.325964] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.326174] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1145.326353] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1145.326580] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1145.326749] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1145.326924] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1145.327134] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1145.327423] env[70013]: DEBUG nova.virt.hardware [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1145.328543] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f461f0-b883-4553-8579-1b4e9ad65b8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.339170] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1fd784-9c5c-4f38-bcf0-43b9d491d8ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.493986] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231519, 'name': ReconfigVM_Task, 'duration_secs': 1.122302} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.494469] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Reconfigured VM instance instance-00000068 to attach disk [datastore2] e11550bf-3284-4ae6-9aa9-34923da589ca/e11550bf-3284-4ae6-9aa9-34923da589ca.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1145.499053] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c33f354-7b16-4eba-b136-99504124cd2b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.508761] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523920b4-3729-7490-c46e-631329266243, 'name': SearchDatastore_Task, 'duration_secs': 0.01078} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.510648] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.510987] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1145.511256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.511411] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.511607] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1145.512029] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1145.512029] env[70013]: value = "task-4231521" [ 1145.512029] env[70013]: _type = "Task" [ 1145.512029] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.512256] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb3858d5-26f5-496f-ac56-e306b669a02f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.525711] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231521, 'name': Rename_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.528352] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1145.528502] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1145.529225] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea4060b1-4766-4f63-8646-7e910c121ec7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.535645] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1145.535645] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a00572-fb5e-0f2e-c62b-74da80187a09" [ 1145.535645] env[70013]: _type = "Task" [ 1145.535645] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.545650] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a00572-fb5e-0f2e-c62b-74da80187a09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.615802] env[70013]: DEBUG oslo_vmware.api [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231520, 'name': PowerOnVM_Task, 'duration_secs': 0.738239} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.616095] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Resumed the VM {{(pid=70013) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1145.616286] env[70013]: DEBUG nova.compute.manager [None req-d8e67524-1fa9-4696-ac06-ccea354fa727 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1145.617081] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6eae238-ca6a-4b75-8ac5-d6808dc14cd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.712660] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d5e328be-8441-4768-a798-53b3d2c2d955 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "499bb115-146c-48dc-9447-3c01c9549a73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.058s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.716939] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.483s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.717166] env[70013]: INFO nova.compute.manager [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Migrating [ 1145.723655] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.167s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1145.723884] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.723990] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1145.724342] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.852s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1145.724576] env[70013]: DEBUG nova.objects.instance [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lazy-loading 'resources' on Instance uuid 8f4d999d-54e9-4987-81f7-27a2b6d4ed17 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.728483] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637c21cf-61ca-41f7-8fbc-1b5c3cd317e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.743733] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa98dee8-e18d-4932-89c1-03e4f199f8ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.760707] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25854159-1292-45b1-85af-a50538f71297 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.769248] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b2cc98-f99f-4bba-9c51-2537556b60ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.804047] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179728MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1145.804289] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1145.891635] env[70013]: DEBUG nova.compute.manager [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Received event network-changed-cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1145.892210] env[70013]: DEBUG nova.compute.manager [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Refreshing instance network info cache due to event network-changed-cc3af4a9-2d59-4b89-975c-87e4a5e6d939. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1145.892624] env[70013]: DEBUG oslo_concurrency.lockutils [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] Acquiring lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.892694] env[70013]: DEBUG oslo_concurrency.lockutils [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] Acquired lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.892856] env[70013]: DEBUG nova.network.neutron [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Refreshing network info cache for port cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1145.935351] env[70013]: DEBUG nova.network.neutron [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Successfully updated port: 43a074aa-e66b-4258-8c74-fae23dca9c71 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1146.025224] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231521, 'name': Rename_Task, 'duration_secs': 0.433874} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.025414] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1146.025594] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6de61c1d-f51c-4044-9177-19f46651c692 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.033560] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1146.033560] env[70013]: value = "task-4231522" [ 1146.033560] env[70013]: _type = "Task" [ 1146.033560] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.045873] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.050088] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a00572-fb5e-0f2e-c62b-74da80187a09, 'name': SearchDatastore_Task, 'duration_secs': 0.017567} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.050940] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00b26be7-414b-4b80-9781-c57c79706c22 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.057200] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1146.057200] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b3bb77-279d-f707-c90a-1da3e01f3af4" [ 1146.057200] env[70013]: _type = "Task" [ 1146.057200] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.067384] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b3bb77-279d-f707-c90a-1da3e01f3af4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.240103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.240283] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.240467] env[70013]: DEBUG nova.network.neutron [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1146.438566] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.438796] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.438956] env[70013]: DEBUG nova.network.neutron [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1146.545203] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231522, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.559617] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72607366-398e-46ce-b6d2-02f3bedc8759 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.575967] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b3bb77-279d-f707-c90a-1da3e01f3af4, 'name': SearchDatastore_Task, 'duration_secs': 0.020299} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.576375] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1146.576643] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 1eea3d47-589d-450b-b5f8-2e763cd1bbd3/1eea3d47-589d-450b-b5f8-2e763cd1bbd3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1146.577684] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c03cf89-0975-4575-a80d-456d04378f58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.581466] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-161e8e36-20e6-4e4a-b3ab-573e1fd9c567 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.615255] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73dbfe27-2601-4961-a21d-3cac8d2be314 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.618068] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1146.618068] env[70013]: value = "task-4231523" [ 1146.618068] env[70013]: _type = "Task" [ 1146.618068] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.625677] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9721a5b2-169a-45d1-aa32-75b945657201 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.633361] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.645665] env[70013]: DEBUG nova.compute.provider_tree [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.851599] env[70013]: DEBUG nova.network.neutron [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Updated VIF entry in instance network info cache for port cc3af4a9-2d59-4b89-975c-87e4a5e6d939. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1146.852072] env[70013]: DEBUG nova.network.neutron [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Updating instance_info_cache with network_info: [{"id": "cc3af4a9-2d59-4b89-975c-87e4a5e6d939", "address": "fa:16:3e:4b:e4:7f", "network": {"id": "8f216d33-a334-43c6-b03f-fc45ad272a21", "bridge": "br-int", "label": "tempest-ServersTestJSON-1533284893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d82869f19af4fbfabc03ecc3c8e59af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc3af4a9-2d", "ovs_interfaceid": "cc3af4a9-2d59-4b89-975c-87e4a5e6d939", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.976093] env[70013]: DEBUG nova.network.neutron [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1147.047972] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231522, 'name': PowerOnVM_Task} progress is 87%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.132866] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231523, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.134281] env[70013]: DEBUG nova.network.neutron [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance_info_cache with network_info: [{"id": "43a074aa-e66b-4258-8c74-fae23dca9c71", "address": "fa:16:3e:2d:02:dc", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a074aa-e6", "ovs_interfaceid": "43a074aa-e66b-4258-8c74-fae23dca9c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.147765] env[70013]: DEBUG nova.scheduler.client.report [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1147.275606] env[70013]: DEBUG nova.network.neutron [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.331970] env[70013]: DEBUG nova.compute.manager [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Received event network-vif-plugged-43a074aa-e66b-4258-8c74-fae23dca9c71 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1147.332208] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] Acquiring lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1147.332472] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.332707] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.332793] env[70013]: DEBUG nova.compute.manager [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] No waiting events found dispatching network-vif-plugged-43a074aa-e66b-4258-8c74-fae23dca9c71 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1147.332944] env[70013]: WARNING nova.compute.manager [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Received unexpected event network-vif-plugged-43a074aa-e66b-4258-8c74-fae23dca9c71 for instance with vm_state building and task_state spawning. [ 1147.333593] env[70013]: DEBUG nova.compute.manager [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Received event network-changed-43a074aa-e66b-4258-8c74-fae23dca9c71 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1147.333789] env[70013]: DEBUG nova.compute.manager [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Refreshing instance network info cache due to event network-changed-43a074aa-e66b-4258-8c74-fae23dca9c71. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1147.333973] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] Acquiring lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.355747] env[70013]: DEBUG oslo_concurrency.lockutils [req-b7d32784-d82b-4d75-b66f-87efe99c981a req-7860a082-5948-4d2c-b326-cf83c5e474e2 service nova] Releasing lock "refresh_cache-499bb115-146c-48dc-9447-3c01c9549a73" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.547077] env[70013]: DEBUG oslo_vmware.api [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231522, 'name': PowerOnVM_Task, 'duration_secs': 1.219716} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.547077] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1147.547077] env[70013]: INFO nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Took 9.36 seconds to spawn the instance on the hypervisor. [ 1147.547077] env[70013]: DEBUG nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1147.547758] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ae0747-a7b9-4738-80b6-26cc062d5e29 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.629088] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.749547} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.629385] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 1eea3d47-589d-450b-b5f8-2e763cd1bbd3/1eea3d47-589d-450b-b5f8-2e763cd1bbd3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1147.629605] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1147.629869] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce8f811e-0aed-4b42-911e-144452c8aebb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.637871] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.638228] env[70013]: DEBUG nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Instance network_info: |[{"id": "43a074aa-e66b-4258-8c74-fae23dca9c71", "address": "fa:16:3e:2d:02:dc", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a074aa-e6", "ovs_interfaceid": "43a074aa-e66b-4258-8c74-fae23dca9c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1147.639816] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] Acquired lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1147.640016] env[70013]: DEBUG nova.network.neutron [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Refreshing network info cache for port 43a074aa-e66b-4258-8c74-fae23dca9c71 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1147.643370] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:02:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43a074aa-e66b-4258-8c74-fae23dca9c71', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1147.652312] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1147.652669] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1147.652669] env[70013]: value = "task-4231524" [ 1147.652669] env[70013]: _type = "Task" [ 1147.652669] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.653739] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1147.654483] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.656714] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71101807-250e-4ada-94a1-6ba16273bd91 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.673925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.395s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.674195] env[70013]: DEBUG nova.objects.instance [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lazy-loading 'resources' on Instance uuid c57e876a-8696-4ce5-b533-7580f484414e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.683188] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.684410] env[70013]: INFO nova.scheduler.client.report [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Deleted allocations for instance 8f4d999d-54e9-4987-81f7-27a2b6d4ed17 [ 1147.686698] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1147.686698] env[70013]: value = "task-4231525" [ 1147.686698] env[70013]: _type = "Task" [ 1147.686698] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.702350] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231525, 'name': CreateVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.779228] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1148.066066] env[70013]: INFO nova.compute.manager [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Took 24.07 seconds to build instance. [ 1148.165927] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.446053} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.165927] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1148.168111] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88abc107-54eb-4836-986c-afea26ba7505 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.193922] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 1eea3d47-589d-450b-b5f8-2e763cd1bbd3/1eea3d47-589d-450b-b5f8-2e763cd1bbd3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1148.194620] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fe9d814-e6e7-47a7-a3d8-4e3f802745a8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.219604] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5b8ef192-c61f-45f3-a144-fffcaabb99fb tempest-ServerMetadataNegativeTestJSON-7641013 tempest-ServerMetadataNegativeTestJSON-7641013-project-member] Lock "8f4d999d-54e9-4987-81f7-27a2b6d4ed17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.399s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.228257] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231525, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.230643] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1148.230643] env[70013]: value = "task-4231526" [ 1148.230643] env[70013]: _type = "Task" [ 1148.230643] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.245876] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231526, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.541406] env[70013]: DEBUG nova.network.neutron [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updated VIF entry in instance network info cache for port 43a074aa-e66b-4258-8c74-fae23dca9c71. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1148.541973] env[70013]: DEBUG nova.network.neutron [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance_info_cache with network_info: [{"id": "43a074aa-e66b-4258-8c74-fae23dca9c71", "address": "fa:16:3e:2d:02:dc", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a074aa-e6", "ovs_interfaceid": "43a074aa-e66b-4258-8c74-fae23dca9c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.561393] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15fb500a-312d-4470-8375-3d813a9e2351 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.569803] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2913cab7-f9c4-42a7-b49c-50d66df3b158 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.589s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.571106] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2a6129-1327-494d-8a06-5476c27fe24d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.607021] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9435157b-e126-4d27-9c07-e4d2541fff39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.618261] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6fe9e2-9eea-4070-a137-b069d79b50bd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.635244] env[70013]: DEBUG nova.compute.provider_tree [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.710651] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231525, 'name': CreateVM_Task, 'duration_secs': 0.98957} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.710950] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1148.712029] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.712308] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1148.712777] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1148.713164] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b28cc90-4a38-4332-a02f-3d7996ea343d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.721107] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1148.721107] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b54fb1-22f1-0593-d0f6-b071af1c907e" [ 1148.721107] env[70013]: _type = "Task" [ 1148.721107] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.732141] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b54fb1-22f1-0593-d0f6-b071af1c907e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.741281] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231526, 'name': ReconfigVM_Task, 'duration_secs': 0.417414} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.742278] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 1eea3d47-589d-450b-b5f8-2e763cd1bbd3/1eea3d47-589d-450b-b5f8-2e763cd1bbd3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.742408] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2feae483-f9c7-4a21-bb3b-280f64fabf56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.749781] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1148.749781] env[70013]: value = "task-4231527" [ 1148.749781] env[70013]: _type = "Task" [ 1148.749781] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.760484] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231527, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.045334] env[70013]: DEBUG oslo_concurrency.lockutils [req-3d19e8ad-7420-450d-a5d5-ca235f7bb6e9 req-d3360541-a524-444a-91ab-90512489d9a8 service nova] Releasing lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1149.140054] env[70013]: DEBUG nova.scheduler.client.report [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1149.233059] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52b54fb1-22f1-0593-d0f6-b071af1c907e, 'name': SearchDatastore_Task, 'duration_secs': 0.032834} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.233393] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1149.233669] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1149.233915] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.234080] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1149.234274] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1149.234557] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ede23513-e291-4aa9-8620-181e17afecc9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.249907] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1149.250156] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1149.254268] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43aa2e8a-c82b-4429-a747-7d211e1cb94b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.260190] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1149.260190] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]522e463b-a266-2836-4004-7ebffa79f023" [ 1149.260190] env[70013]: _type = "Task" [ 1149.260190] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.263713] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231527, 'name': Rename_Task, 'duration_secs': 0.430529} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.267322] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1149.267615] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1903c3b5-3f88-4776-9b81-4f36b0ddca97 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.278038] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1149.278038] env[70013]: value = "task-4231528" [ 1149.278038] env[70013]: _type = "Task" [ 1149.278038] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.278336] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522e463b-a266-2836-4004-7ebffa79f023, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.286740] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231528, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.294769] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3cde220-b6d0-4e82-ba11-ab890eb64c15 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.314756] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance '28a56a75-9b30-4121-8252-a9e57287441c' progress to 0 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1149.372974] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "e11550bf-3284-4ae6-9aa9-34923da589ca" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.372974] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.373209] env[70013]: DEBUG nova.compute.manager [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1149.375773] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a715d1b-669e-4a77-91d1-0404e761f379 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.387275] env[70013]: DEBUG nova.compute.manager [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1149.387275] env[70013]: DEBUG nova.objects.instance [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'flavor' on Instance uuid e11550bf-3284-4ae6-9aa9-34923da589ca {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.646478] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.971s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.647467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.843s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.686994] env[70013]: INFO nova.scheduler.client.report [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Deleted allocations for instance c57e876a-8696-4ce5-b533-7580f484414e [ 1149.780603] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]522e463b-a266-2836-4004-7ebffa79f023, 'name': SearchDatastore_Task, 'duration_secs': 0.032162} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.786603] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c098dc9d-3724-4f78-8eaa-b0101dd0991a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.795326] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231528, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.796735] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1149.796735] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]520073a1-816e-dda4-2805-22b0cd276df2" [ 1149.796735] env[70013]: _type = "Task" [ 1149.796735] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.806124] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520073a1-816e-dda4-2805-22b0cd276df2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.821021] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1149.821330] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d5006f2-318f-4964-9728-8eda35f6877c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.829684] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1149.829684] env[70013]: value = "task-4231529" [ 1149.829684] env[70013]: _type = "Task" [ 1149.829684] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.839283] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.197850] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8d681da8-7f2e-4a60-b42a-6e809f14c81c tempest-ServerRescueTestJSON-144377210 tempest-ServerRescueTestJSON-144377210-project-member] Lock "c57e876a-8696-4ce5-b533-7580f484414e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.459s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1150.290864] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231528, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.306683] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520073a1-816e-dda4-2805-22b0cd276df2, 'name': SearchDatastore_Task, 'duration_secs': 0.034929} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.306992] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1150.307272] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8/dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1150.307535] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24dbf7bb-50b5-4a05-8010-29ef936c64b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.315390] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1150.315390] env[70013]: value = "task-4231530" [ 1150.315390] env[70013]: _type = "Task" [ 1150.315390] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.324445] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231530, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.340583] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231529, 'name': PowerOffVM_Task, 'duration_secs': 0.388249} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.340583] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1150.341018] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance '28a56a75-9b30-4121-8252-a9e57287441c' progress to 17 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1150.398303] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1150.398621] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f986830-88e5-470a-842a-e692d90698f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.407113] env[70013]: DEBUG oslo_vmware.api [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1150.407113] env[70013]: value = "task-4231531" [ 1150.407113] env[70013]: _type = "Task" [ 1150.407113] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.417552] env[70013]: DEBUG oslo_vmware.api [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.663987] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Applying migration context for instance 28a56a75-9b30-4121-8252-a9e57287441c as it has an incoming, in-progress migration 8f1251c5-4dcb-4b6f-afa5-81f973aaec05. Migration status is migrating {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1150.668660] env[70013]: INFO nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating resource usage from migration 8f1251c5-4dcb-4b6f-afa5-81f973aaec05 [ 1150.696708] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance fc473d40-b57b-437e-9511-58a0ba700a69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.696886] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.696991] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 09544922-6dc4-48b5-8cfd-e91e7f74c13f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.697180] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.697385] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.697439] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.697556] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.697724] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.697867] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.697867] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698029] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9026841c-7bda-41e5-a4ac-03d0d3e37560 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698073] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 472b8218-51c3-492e-96ff-5fa99df4cc4c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698191] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 38738d80-25ca-4a17-81f1-e0a7bc8c4a44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698297] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 0ea84d0d-905c-428d-8abb-2781c817f08f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698947] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 499bb115-146c-48dc-9447-3c01c9549a73 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698947] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance e11550bf-3284-4ae6-9aa9-34923da589ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698947] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 1eea3d47-589d-450b-b5f8-2e763cd1bbd3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698947] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.698947] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Migration 8f1251c5-4dcb-4b6f-afa5-81f973aaec05 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1742}} [ 1150.698947] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 28a56a75-9b30-4121-8252-a9e57287441c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1150.699512] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 20 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1150.699512] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4480MB phys_disk=200GB used_disk=20GB total_vcpus=48 used_vcpus=20 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1150.793963] env[70013]: DEBUG oslo_vmware.api [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231528, 'name': PowerOnVM_Task, 'duration_secs': 1.247231} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.794270] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1150.794479] env[70013]: INFO nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Took 10.10 seconds to spawn the instance on the hypervisor. [ 1150.794664] env[70013]: DEBUG nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1150.795493] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e110a647-922d-47e8-a46e-ef91aad3b59e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.829657] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231530, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.850115] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1150.852285] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1150.857427] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c1c05d7-9666-4f7e-85ee-b41642114e81 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.880063] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1150.880063] env[70013]: value = "task-4231532" [ 1150.880063] env[70013]: _type = "Task" [ 1150.880063] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.892384] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231532, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.931677] env[70013]: DEBUG oslo_vmware.api [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231531, 'name': PowerOffVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.074725] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63d23eb-8713-465e-8b53-b772fa126d98 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.083313] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827b6762-3c22-44aa-88bb-36c941508616 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.122064] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8cb5e8-fa05-4b44-bc69-e6d8ff607105 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.133431] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63542b9d-e5d9-4283-b415-351ba964b94d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.135846] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.136140] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.136349] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.136532] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.136705] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.139290] env[70013]: INFO nova.compute.manager [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Terminating instance [ 1151.152334] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.317238] env[70013]: INFO nova.compute.manager [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Took 25.64 seconds to build instance. [ 1151.330440] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231530, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.391728] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231532, 'name': ReconfigVM_Task, 'duration_secs': 0.408766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.392964] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance '28a56a75-9b30-4121-8252-a9e57287441c' progress to 33 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1151.422810] env[70013]: DEBUG oslo_vmware.api [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231531, 'name': PowerOffVM_Task, 'duration_secs': 0.612155} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.423255] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1151.423434] env[70013]: DEBUG nova.compute.manager [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1151.424446] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4c67ab-53d2-4545-af5b-340459295783 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.657907] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1151.661775] env[70013]: DEBUG nova.compute.manager [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1151.661988] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1151.663407] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7134778-24e4-4b29-b307-9f5014f491f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.672870] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1151.673207] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf473c40-30c6-42e8-9cc2-d717c250f8c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.683394] env[70013]: DEBUG oslo_vmware.api [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1151.683394] env[70013]: value = "task-4231533" [ 1151.683394] env[70013]: _type = "Task" [ 1151.683394] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.695764] env[70013]: DEBUG oslo_vmware.api [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.819639] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7b240801-af8e-459d-a6b7-9a12a33bf89e tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.159s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.830643] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231530, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.322719} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.830967] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8/dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1151.831180] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1151.831439] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7cfdc24b-486c-4285-9d90-85ae225d868a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.839945] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1151.839945] env[70013]: value = "task-4231534" [ 1151.839945] env[70013]: _type = "Task" [ 1151.839945] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.848588] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231534, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.904545] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1151.904834] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1151.904996] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1151.905288] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1151.905449] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1151.905602] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1151.905811] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1151.905988] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1151.906192] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1151.906361] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1151.906536] env[70013]: DEBUG nova.virt.hardware [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1151.912698] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Reconfiguring VM instance instance-00000009 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1151.913112] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a59c7f1-b379-4a48-9022-1750c3483bd5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.938000] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1151.938000] env[70013]: value = "task-4231535" [ 1151.938000] env[70013]: _type = "Task" [ 1151.938000] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.938697] env[70013]: DEBUG oslo_concurrency.lockutils [None req-9330441f-d8c5-4a4e-b09d-934dc0ea7b77 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.566s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.958081] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231535, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.163491] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1152.164037] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.516s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.197801] env[70013]: DEBUG oslo_vmware.api [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231533, 'name': PowerOffVM_Task, 'duration_secs': 0.367183} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.197801] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1152.198028] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1152.198274] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e784c0d-b832-4f68-b974-466d8b3ab46b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.281948] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1152.282229] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1152.282419] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleting the datastore file [datastore1] 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1152.282802] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95c8dd74-4482-4b2e-b8bf-22a1d1e9a55a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.292956] env[70013]: DEBUG oslo_vmware.api [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for the task: (returnval){ [ 1152.292956] env[70013]: value = "task-4231537" [ 1152.292956] env[70013]: _type = "Task" [ 1152.292956] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.303957] env[70013]: DEBUG oslo_vmware.api [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.350279] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231534, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.269089} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.350783] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1152.351727] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df65b5b-e05c-4118-8cef-8e22860a3db5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.382323] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8/dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.382323] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e19cffe-2cfc-48f8-98c8-86ce7db7d3e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.403546] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1152.403546] env[70013]: value = "task-4231538" [ 1152.403546] env[70013]: _type = "Task" [ 1152.403546] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.415063] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.452117] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231535, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.811494] env[70013]: DEBUG oslo_vmware.api [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Task: {'id': task-4231537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.434417} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.811494] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1152.811494] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1152.811494] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1152.811494] env[70013]: INFO nova.compute.manager [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1152.811494] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1152.811494] env[70013]: DEBUG nova.compute.manager [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1152.811494] env[70013]: DEBUG nova.network.neutron [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1152.917946] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.951964] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231535, 'name': ReconfigVM_Task, 'duration_secs': 0.928288} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.952312] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Reconfigured VM instance instance-00000009 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1152.953240] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29df98f-4d26-483e-accd-1f6e69ebb767 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.978713] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] 28a56a75-9b30-4121-8252-a9e57287441c/28a56a75-9b30-4121-8252-a9e57287441c.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.979142] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4559ef73-ad68-4b85-8661-d2e7bd57ac31 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.000350] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1153.000350] env[70013]: value = "task-4231539" [ 1153.000350] env[70013]: _type = "Task" [ 1153.000350] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.010029] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.416277] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.511857] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231539, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.561590] env[70013]: DEBUG nova.compute.manager [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1153.561590] env[70013]: DEBUG nova.compute.manager [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing instance network info cache due to event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1153.561590] env[70013]: DEBUG oslo_concurrency.lockutils [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.561590] env[70013]: DEBUG oslo_concurrency.lockutils [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.561823] env[70013]: DEBUG nova.network.neutron [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1153.916222] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231538, 'name': ReconfigVM_Task, 'duration_secs': 1.233441} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.916578] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Reconfigured VM instance instance-0000006a to attach disk [datastore2] dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8/dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1153.917218] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30d2d41e-5db2-476f-8413-05b0ce0496fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.926675] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1153.926675] env[70013]: value = "task-4231540" [ 1153.926675] env[70013]: _type = "Task" [ 1153.926675] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.939473] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231540, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.014390] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231539, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.032936] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "e11550bf-3284-4ae6-9aa9-34923da589ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.032936] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.032936] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "e11550bf-3284-4ae6-9aa9-34923da589ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.032936] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.032936] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.034358] env[70013]: INFO nova.compute.manager [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Terminating instance [ 1154.300630] env[70013]: DEBUG nova.network.neutron [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.442935] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231540, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.512753] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231539, 'name': ReconfigVM_Task, 'duration_secs': 1.025169} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.513120] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Reconfigured VM instance instance-00000009 to attach disk [datastore2] 28a56a75-9b30-4121-8252-a9e57287441c/28a56a75-9b30-4121-8252-a9e57287441c.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1154.513422] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance '28a56a75-9b30-4121-8252-a9e57287441c' progress to 50 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1154.531781] env[70013]: DEBUG nova.network.neutron [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updated VIF entry in instance network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1154.532165] env[70013]: DEBUG nova.network.neutron [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.539751] env[70013]: DEBUG nova.compute.manager [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1154.539986] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1154.540979] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705fc8e4-0195-46a1-8bbb-bebf08dc1af1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.549680] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1154.549947] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-edc502e6-1149-4f97-877d-78df59071c8a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.653195] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1154.653532] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1154.653742] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleting the datastore file [datastore2] e11550bf-3284-4ae6-9aa9-34923da589ca {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.654053] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae5f14ec-0fb6-4560-b6d6-c78d045806ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.661347] env[70013]: DEBUG oslo_vmware.api [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1154.661347] env[70013]: value = "task-4231542" [ 1154.661347] env[70013]: _type = "Task" [ 1154.661347] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.670527] env[70013]: DEBUG oslo_vmware.api [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231542, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.803373] env[70013]: INFO nova.compute.manager [-] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Took 1.99 seconds to deallocate network for instance. [ 1154.940330] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231540, 'name': Rename_Task, 'duration_secs': 0.542872} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.940698] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1154.940910] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-febebeeb-e00b-444c-8bdd-d875734e8f2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.948597] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1154.948597] env[70013]: value = "task-4231543" [ 1154.948597] env[70013]: _type = "Task" [ 1154.948597] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.958058] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.020785] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d610e29d-6df8-44dc-bd8b-835f39e912ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.043104] env[70013]: DEBUG oslo_concurrency.lockutils [req-91862e55-16a4-4106-b8e2-e511334b187a req-5df2977e-dfbc-4a0d-93d5-a72e72ee28e7 service nova] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.044290] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb4473f-b10c-462f-be97-954546545cab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.063647] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance '28a56a75-9b30-4121-8252-a9e57287441c' progress to 67 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1155.172484] env[70013]: DEBUG oslo_vmware.api [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231542, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.319323] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1155.319611] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.319857] env[70013]: DEBUG nova.objects.instance [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lazy-loading 'resources' on Instance uuid 0ea84d0d-905c-428d-8abb-2781c817f08f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.460287] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231543, 'name': PowerOnVM_Task} progress is 76%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.592245] env[70013]: DEBUG nova.compute.manager [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1155.592414] env[70013]: DEBUG nova.compute.manager [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing instance network info cache due to event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1155.592732] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.592859] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.592953] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1155.614390] env[70013]: DEBUG nova.network.neutron [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Port a0d11944-d177-42a5-ad2b-22c116396f8a binding to destination host cpu-1 is already ACTIVE {{(pid=70013) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1155.672944] env[70013]: DEBUG oslo_vmware.api [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231542, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.601908} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.673253] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.673455] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1155.673903] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1155.674101] env[70013]: INFO nova.compute.manager [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1155.674351] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1155.674555] env[70013]: DEBUG nova.compute.manager [-] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1155.674655] env[70013]: DEBUG nova.network.neutron [-] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1155.965553] env[70013]: DEBUG oslo_vmware.api [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231543, 'name': PowerOnVM_Task, 'duration_secs': 1.008822} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.965553] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1155.965888] env[70013]: INFO nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Took 10.70 seconds to spawn the instance on the hypervisor. [ 1155.966045] env[70013]: DEBUG nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1155.966807] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc804832-d40f-4c9e-bacd-1e8e515e273b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.132163] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77e9976-fa40-4bd6-8d7f-14af303c3019 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.143225] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3eaa454-29ad-4466-bfba-f1d192cbec49 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.180101] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e5d560-38e7-4b4d-ac5e-75295e76dc38 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.190778] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1bf78d-6623-43a1-a63c-2a804b249462 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.206087] env[70013]: DEBUG nova.compute.provider_tree [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.322363] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updated VIF entry in instance network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1156.322790] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.462668] env[70013]: DEBUG nova.network.neutron [-] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.487021] env[70013]: INFO nova.compute.manager [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Took 25.30 seconds to build instance. [ 1156.638838] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.638838] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.638838] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.711345] env[70013]: DEBUG nova.scheduler.client.report [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1156.826117] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1156.826117] env[70013]: DEBUG nova.compute.manager [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Received event network-vif-deleted-d8aacd8d-2fc2-477b-9d82-e101e20b0132 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1156.826117] env[70013]: DEBUG nova.compute.manager [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1156.826117] env[70013]: DEBUG nova.compute.manager [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing instance network info cache due to event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1156.826354] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.826354] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.826467] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1156.965344] env[70013]: INFO nova.compute.manager [-] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Took 1.29 seconds to deallocate network for instance. [ 1156.988783] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ba13453e-ee86-4ebd-8ae2-6e4918c68004 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.815s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.216024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.896s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.239375] env[70013]: INFO nova.scheduler.client.report [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Deleted allocations for instance 0ea84d0d-905c-428d-8abb-2781c817f08f [ 1157.472062] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.472492] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.472814] env[70013]: DEBUG nova.objects.instance [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'resources' on Instance uuid e11550bf-3284-4ae6-9aa9-34923da589ca {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.580381] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updated VIF entry in instance network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1157.580764] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.625484] env[70013]: DEBUG nova.compute.manager [req-61f220cb-9061-4b6f-8ad6-f0f776f05b00 req-66b8360b-a2c7-44fd-9d31-8f22137b8775 service nova] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Received event network-vif-deleted-3eb52fcc-22d7-430e-9fd7-54b925e05cec {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1157.695177] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.695376] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1157.695559] env[70013]: DEBUG nova.network.neutron [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1157.748476] env[70013]: DEBUG oslo_concurrency.lockutils [None req-64258412-f889-4b0c-bfa7-21a2a168e999 tempest-ServersNegativeTestJSON-161729029 tempest-ServersNegativeTestJSON-161729029-project-member] Lock "0ea84d0d-905c-428d-8abb-2781c817f08f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.612s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.083935] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1158.084254] env[70013]: DEBUG nova.compute.manager [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1158.084455] env[70013]: DEBUG nova.compute.manager [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing instance network info cache due to event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1158.084681] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.084902] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.084979] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1158.245848] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2d9bca-3a42-4234-9c11-3fb520da05ee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.254880] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f89b18f-8c79-4b3d-a896-a09968edf70b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.304154] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e946def-864f-480f-9188-325a3081dca4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.311615] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5351f227-367b-4940-a2b8-ebdb8bbfbe08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.326239] env[70013]: DEBUG nova.compute.provider_tree [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.599666] env[70013]: DEBUG nova.compute.manager [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Stashing vm_state: active {{(pid=70013) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1158.606108] env[70013]: DEBUG nova.network.neutron [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.829200] env[70013]: DEBUG nova.scheduler.client.report [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1158.966735] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updated VIF entry in instance network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1158.966920] env[70013]: DEBUG nova.network.neutron [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.111272] env[70013]: DEBUG oslo_concurrency.lockutils [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.132753] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.334870] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.337295] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.205s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.362614] env[70013]: INFO nova.scheduler.client.report [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted allocations for instance e11550bf-3284-4ae6-9aa9-34923da589ca [ 1159.469968] env[70013]: DEBUG oslo_concurrency.lockutils [req-cd61a700-9778-497b-bb3d-205499db2fee req-16000286-0626-4ffa-ae6d-8c4525e8389c service nova] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.578888] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.579186] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.635442] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3a0fa3-dd47-4889-99e5-5efcd6671069 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.657025] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b0614a-9777-4f21-bc5f-087a54c36123 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.664735] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance '28a56a75-9b30-4121-8252-a9e57287441c' progress to 83 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1159.843360] env[70013]: INFO nova.compute.claims [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1159.873379] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b913a56e-874d-4cd7-9cdb-5a7b97334335 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "e11550bf-3284-4ae6-9aa9-34923da589ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.843s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1160.082378] env[70013]: INFO nova.compute.manager [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Detaching volume 5a787b14-e921-4a07-a8c6-06fd861030f1 [ 1160.119383] env[70013]: INFO nova.virt.block_device [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Attempting to driver detach volume 5a787b14-e921-4a07-a8c6-06fd861030f1 from mountpoint /dev/sdb [ 1160.119764] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1160.119818] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837209', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'name': 'volume-5a787b14-e921-4a07-a8c6-06fd861030f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '472b8218-51c3-492e-96ff-5fa99df4cc4c', 'attached_at': '', 'detached_at': '', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'serial': '5a787b14-e921-4a07-a8c6-06fd861030f1'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1160.120737] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61f4549-2dec-463a-ac28-6ecf0dbb84b4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.142570] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88850ce-90be-4c86-89c0-b299d00a388b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.150528] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16933b17-2c20-44de-861a-320a4b28d3db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.172800] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1160.173201] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f53c5da6-eba6-44a6-807d-581b9813743a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.175566] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa364ce-f164-4f50-902b-aa6fc608565e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.193603] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] The volume has not been displaced from its original location: [datastore1] volume-5a787b14-e921-4a07-a8c6-06fd861030f1/volume-5a787b14-e921-4a07-a8c6-06fd861030f1.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1160.199330] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1160.201147] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66307a44-57ed-4527-8d28-2164c284b458 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.214506] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1160.214506] env[70013]: value = "task-4231544" [ 1160.214506] env[70013]: _type = "Task" [ 1160.214506] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.220977] env[70013]: DEBUG oslo_vmware.api [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1160.220977] env[70013]: value = "task-4231545" [ 1160.220977] env[70013]: _type = "Task" [ 1160.220977] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.224334] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231544, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.232298] env[70013]: DEBUG oslo_vmware.api [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231545, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.351666] env[70013]: INFO nova.compute.resource_tracker [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating resource usage from migration 2d43898c-e276-4439-863b-6de02d09ffa7 [ 1160.618770] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f9ef6b-2185-44cd-9105-5bf1d96fd356 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.628326] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ced3573-bec4-4d4d-99af-01c11bad2cce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.661996] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c7e61b-f1dd-4603-bdef-0b5209680e65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.670378] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e7c3c0-6c86-4cd6-8a5b-28e3b7cfbdf3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.685512] env[70013]: DEBUG nova.compute.provider_tree [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.725211] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231544, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.734330] env[70013]: DEBUG oslo_vmware.api [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231545, 'name': ReconfigVM_Task, 'duration_secs': 0.253203} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.734617] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1160.739613] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d83c8f4f-0e0c-4032-ae48-127547428b63 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.755479] env[70013]: DEBUG oslo_vmware.api [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1160.755479] env[70013]: value = "task-4231546" [ 1160.755479] env[70013]: _type = "Task" [ 1160.755479] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.767077] env[70013]: DEBUG oslo_vmware.api [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231546, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.024353] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.024520] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.025066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.025162] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.025357] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.027809] env[70013]: INFO nova.compute.manager [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Terminating instance [ 1161.190039] env[70013]: DEBUG nova.scheduler.client.report [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1161.224682] env[70013]: DEBUG oslo_vmware.api [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231544, 'name': PowerOnVM_Task, 'duration_secs': 0.763448} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.224908] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1161.225104] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-56fa8153-9d7b-4412-9be0-b3590358d656 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance '28a56a75-9b30-4121-8252-a9e57287441c' progress to 100 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1161.266282] env[70013]: DEBUG oslo_vmware.api [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231546, 'name': ReconfigVM_Task, 'duration_secs': 0.365315} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.266644] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837209', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'name': 'volume-5a787b14-e921-4a07-a8c6-06fd861030f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '472b8218-51c3-492e-96ff-5fa99df4cc4c', 'attached_at': '', 'detached_at': '', 'volume_id': '5a787b14-e921-4a07-a8c6-06fd861030f1', 'serial': '5a787b14-e921-4a07-a8c6-06fd861030f1'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1161.531686] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.531957] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquired lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1161.532155] env[70013]: DEBUG nova.network.neutron [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1161.565947] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-38738d80-25ca-4a17-81f1-e0a7bc8c4a44-b209fd87-7748-481e-8262-5ec43c4a436b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.566228] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-38738d80-25ca-4a17-81f1-e0a7bc8c4a44-b209fd87-7748-481e-8262-5ec43c4a436b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.566592] env[70013]: DEBUG nova.objects.instance [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'flavor' on Instance uuid 38738d80-25ca-4a17-81f1-e0a7bc8c4a44 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.698057] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.360s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.698364] env[70013]: INFO nova.compute.manager [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Migrating [ 1161.813083] env[70013]: DEBUG nova.objects.instance [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'flavor' on Instance uuid 472b8218-51c3-492e-96ff-5fa99df4cc4c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.051324] env[70013]: DEBUG nova.network.neutron [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1162.119903] env[70013]: DEBUG nova.network.neutron [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.186656] env[70013]: DEBUG nova.objects.instance [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'pci_requests' on Instance uuid 38738d80-25ca-4a17-81f1-e0a7bc8c4a44 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.214874] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.215187] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.215298] env[70013]: DEBUG nova.network.neutron [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1162.359371] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.359761] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.359998] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.360237] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.360451] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.362521] env[70013]: INFO nova.compute.manager [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Terminating instance [ 1162.622931] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Releasing lock "refresh_cache-9026841c-7bda-41e5-a4ac-03d0d3e37560" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1162.623352] env[70013]: DEBUG nova.compute.manager [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1162.623551] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1162.623881] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f34d19c-fd2b-46bc-8220-418f58930fdc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.634573] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7c4f67-c7c4-4ce2-b6a1-7949ea60802c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.662929] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9026841c-7bda-41e5-a4ac-03d0d3e37560 could not be found. [ 1162.663167] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1162.663351] env[70013]: INFO nova.compute.manager [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1162.663595] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1162.663847] env[70013]: DEBUG nova.compute.manager [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1162.663947] env[70013]: DEBUG nova.network.neutron [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1162.680344] env[70013]: DEBUG nova.network.neutron [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1162.689559] env[70013]: DEBUG nova.objects.base [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Object Instance<38738d80-25ca-4a17-81f1-e0a7bc8c4a44> lazy-loaded attributes: flavor,pci_requests {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1162.689739] env[70013]: DEBUG nova.network.neutron [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1162.755022] env[70013]: DEBUG nova.policy [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1162.820632] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49422425-acba-46d9-86fa-23cad38325f2 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.241s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.865809] env[70013]: DEBUG nova.compute.manager [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1162.866060] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1162.867743] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81229028-81a9-421c-81f4-fa83cf6787aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.876960] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1162.877181] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39cb265a-f8d9-42e0-94ba-68f492c0b89b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.883731] env[70013]: DEBUG oslo_vmware.api [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1162.883731] env[70013]: value = "task-4231547" [ 1162.883731] env[70013]: _type = "Task" [ 1162.883731] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.895589] env[70013]: DEBUG oslo_vmware.api [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.926367] env[70013]: DEBUG nova.network.neutron [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance_info_cache with network_info: [{"id": "43a074aa-e66b-4258-8c74-fae23dca9c71", "address": "fa:16:3e:2d:02:dc", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a074aa-e6", "ovs_interfaceid": "43a074aa-e66b-4258-8c74-fae23dca9c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.183081] env[70013]: DEBUG nova.network.neutron [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.396692] env[70013]: DEBUG oslo_vmware.api [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231547, 'name': PowerOffVM_Task, 'duration_secs': 0.207727} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.397040] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1163.397126] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1163.397386] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80c2d8c6-2d45-4202-88ae-9c1cacabd5eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.429416] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.474264] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1163.474513] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1163.474934] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleting the datastore file [datastore2] 09544922-6dc4-48b5-8cfd-e91e7f74c13f {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.475332] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c89f0bd-f907-4e5a-bd8a-66a9eb11f50b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.483629] env[70013]: DEBUG oslo_vmware.api [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for the task: (returnval){ [ 1163.483629] env[70013]: value = "task-4231549" [ 1163.483629] env[70013]: _type = "Task" [ 1163.483629] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.492525] env[70013]: DEBUG oslo_vmware.api [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.685694] env[70013]: INFO nova.compute.manager [-] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Took 1.02 seconds to deallocate network for instance. [ 1163.929117] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.929375] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.929563] env[70013]: DEBUG nova.compute.manager [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Going to confirm migration 2 {{(pid=70013) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1163.996034] env[70013]: DEBUG oslo_vmware.api [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Task: {'id': task-4231549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160208} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.996337] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1163.996526] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1163.996737] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1163.996938] env[70013]: INFO nova.compute.manager [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1163.997326] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1163.997545] env[70013]: DEBUG nova.compute.manager [-] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1163.997668] env[70013]: DEBUG nova.network.neutron [-] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1164.004035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.005472] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.005472] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.005472] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.005472] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.006898] env[70013]: INFO nova.compute.manager [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Terminating instance [ 1164.193718] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.194375] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.194678] env[70013]: DEBUG nova.objects.instance [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'resources' on Instance uuid 9026841c-7bda-41e5-a4ac-03d0d3e37560 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.277301] env[70013]: DEBUG nova.compute.manager [req-311ae858-a9b1-4b98-9de6-a1ffb5a73d83 req-c6dffdcb-d3d3-4156-9fd6-4c00bc48fc66 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-vif-plugged-b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1164.277983] env[70013]: DEBUG oslo_concurrency.lockutils [req-311ae858-a9b1-4b98-9de6-a1ffb5a73d83 req-c6dffdcb-d3d3-4156-9fd6-4c00bc48fc66 service nova] Acquiring lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.278781] env[70013]: DEBUG oslo_concurrency.lockutils [req-311ae858-a9b1-4b98-9de6-a1ffb5a73d83 req-c6dffdcb-d3d3-4156-9fd6-4c00bc48fc66 service nova] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.279327] env[70013]: DEBUG oslo_concurrency.lockutils [req-311ae858-a9b1-4b98-9de6-a1ffb5a73d83 req-c6dffdcb-d3d3-4156-9fd6-4c00bc48fc66 service nova] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.279832] env[70013]: DEBUG nova.compute.manager [req-311ae858-a9b1-4b98-9de6-a1ffb5a73d83 req-c6dffdcb-d3d3-4156-9fd6-4c00bc48fc66 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] No waiting events found dispatching network-vif-plugged-b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1164.280700] env[70013]: WARNING nova.compute.manager [req-311ae858-a9b1-4b98-9de6-a1ffb5a73d83 req-c6dffdcb-d3d3-4156-9fd6-4c00bc48fc66 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received unexpected event network-vif-plugged-b209fd87-7748-481e-8262-5ec43c4a436b for instance with vm_state active and task_state None. [ 1164.289421] env[70013]: DEBUG nova.network.neutron [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Successfully updated port: b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1164.386547] env[70013]: DEBUG nova.compute.manager [req-9e957b88-da91-482d-a06b-e902c267a53b req-006756b6-4f3d-4173-b5a9-94d2ecb5959c service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Received event network-vif-deleted-3f3abd39-894f-40ce-bea2-c815ae8ed3c3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1164.386758] env[70013]: INFO nova.compute.manager [req-9e957b88-da91-482d-a06b-e902c267a53b req-006756b6-4f3d-4173-b5a9-94d2ecb5959c service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Neutron deleted interface 3f3abd39-894f-40ce-bea2-c815ae8ed3c3; detaching it from the instance and deleting it from the info cache [ 1164.386934] env[70013]: DEBUG nova.network.neutron [req-9e957b88-da91-482d-a06b-e902c267a53b req-006756b6-4f3d-4173-b5a9-94d2ecb5959c service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.511115] env[70013]: DEBUG nova.compute.manager [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1164.511458] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1164.512393] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e100eb-06ae-4aab-8c70-a877c813f2e3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.522298] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1164.522687] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d883509f-3269-4b6b-af88-dd07d8426a95 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.530592] env[70013]: DEBUG oslo_vmware.api [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1164.530592] env[70013]: value = "task-4231550" [ 1164.530592] env[70013]: _type = "Task" [ 1164.530592] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.540726] env[70013]: DEBUG oslo_vmware.api [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231550, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.748837] env[70013]: DEBUG nova.network.neutron [-] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.783153] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.783346] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.783521] env[70013]: DEBUG nova.network.neutron [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1164.783751] env[70013]: DEBUG nova.objects.instance [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'info_cache' on Instance uuid 28a56a75-9b30-4121-8252-a9e57287441c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.792539] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.792730] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.792917] env[70013]: DEBUG nova.network.neutron [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1164.889294] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0475140d-ed18-4629-87a0-c4f5e54add41 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.901150] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfc7faa-51b3-4a85-aa1b-1fce697536f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.936793] env[70013]: DEBUG nova.compute.manager [req-9e957b88-da91-482d-a06b-e902c267a53b req-006756b6-4f3d-4173-b5a9-94d2ecb5959c service nova] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Detach interface failed, port_id=3f3abd39-894f-40ce-bea2-c815ae8ed3c3, reason: Instance 09544922-6dc4-48b5-8cfd-e91e7f74c13f could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1164.945935] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d88832-5043-4aeb-893a-8be7c5e74dcd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.969594] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance 'dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8' progress to 0 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1165.011018] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a9c06b-16d2-4a2c-b86b-26ad53d28062 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.019752] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86ec2c7-73be-4ac6-b108-4588cfa86c29 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.055414] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41d9534-cf9a-48af-89a7-5fa28d25b13f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.063425] env[70013]: DEBUG oslo_vmware.api [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231550, 'name': PowerOffVM_Task, 'duration_secs': 0.206292} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.065543] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1165.065725] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1165.066017] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a26823d6-9577-45ca-ab1a-5bf856600e30 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.068322] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82699978-42b7-4869-99bd-825345ebc0bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.082397] env[70013]: DEBUG nova.compute.provider_tree [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.132205] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1165.132443] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1165.132682] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleting the datastore file [datastore1] 472b8218-51c3-492e-96ff-5fa99df4cc4c {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1165.133091] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f40be7f-6046-4c53-b93d-a17bea47a327 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.141148] env[70013]: DEBUG oslo_vmware.api [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1165.141148] env[70013]: value = "task-4231552" [ 1165.141148] env[70013]: _type = "Task" [ 1165.141148] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.151027] env[70013]: DEBUG oslo_vmware.api [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231552, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.254702] env[70013]: INFO nova.compute.manager [-] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Took 1.26 seconds to deallocate network for instance. [ 1165.346081] env[70013]: WARNING nova.network.neutron [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] b160f70a-0663-495b-bd35-440c047157c8 already exists in list: networks containing: ['b160f70a-0663-495b-bd35-440c047157c8']. ignoring it [ 1165.476570] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.476938] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1de25d9e-62b1-49b0-9e35-2e73df88f5bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.485021] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1165.485021] env[70013]: value = "task-4231553" [ 1165.485021] env[70013]: _type = "Task" [ 1165.485021] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.495182] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.586249] env[70013]: DEBUG nova.scheduler.client.report [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1165.652497] env[70013]: DEBUG oslo_vmware.api [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231552, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14828} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.652808] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1165.653010] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1165.653210] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1165.653443] env[70013]: INFO nova.compute.manager [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1165.653780] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1165.654023] env[70013]: DEBUG nova.compute.manager [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1165.654124] env[70013]: DEBUG nova.network.neutron [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1165.695082] env[70013]: DEBUG nova.network.neutron [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b209fd87-7748-481e-8262-5ec43c4a436b", "address": "fa:16:3e:08:bd:72", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb209fd87-77", "ovs_interfaceid": "b209fd87-7748-481e-8262-5ec43c4a436b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.761079] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1165.997798] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231553, 'name': PowerOffVM_Task, 'duration_secs': 0.261258} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.998078] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1165.998261] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance 'dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8' progress to 17 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1166.068121] env[70013]: DEBUG nova.network.neutron [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [{"id": "a0d11944-d177-42a5-ad2b-22c116396f8a", "address": "fa:16:3e:6c:10:44", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0d11944-d1", "ovs_interfaceid": "a0d11944-d177-42a5-ad2b-22c116396f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.094562] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.097131] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.336s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.097995] env[70013]: DEBUG nova.objects.instance [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lazy-loading 'resources' on Instance uuid 09544922-6dc4-48b5-8cfd-e91e7f74c13f {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.127371] env[70013]: INFO nova.scheduler.client.report [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted allocations for instance 9026841c-7bda-41e5-a4ac-03d0d3e37560 [ 1166.198022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.198410] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.198563] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.199733] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf34c05b-0bd6-4842-987a-deefce3d42ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.219802] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1166.219802] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1166.219802] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1166.220107] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1166.220107] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1166.220335] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1166.220448] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1166.220726] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1166.220863] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1166.221050] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1166.221231] env[70013]: DEBUG nova.virt.hardware [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1166.227724] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Reconfiguring VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1166.228098] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7daff61b-1ba5-4277-9857-78008f8e33bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.248659] env[70013]: DEBUG oslo_vmware.api [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1166.248659] env[70013]: value = "task-4231554" [ 1166.248659] env[70013]: _type = "Task" [ 1166.248659] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.258419] env[70013]: DEBUG oslo_vmware.api [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231554, 'name': ReconfigVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.418858] env[70013]: DEBUG nova.compute.manager [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-changed-b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1166.419109] env[70013]: DEBUG nova.compute.manager [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing instance network info cache due to event network-changed-b209fd87-7748-481e-8262-5ec43c4a436b. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1166.419351] env[70013]: DEBUG oslo_concurrency.lockutils [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.419499] env[70013]: DEBUG oslo_concurrency.lockutils [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.419663] env[70013]: DEBUG nova.network.neutron [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing network info cache for port b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1166.491269] env[70013]: DEBUG nova.compute.manager [req-a0b6c14b-24f7-4288-bd18-9c4c685171cc req-e732f3f1-112e-4915-9bba-b0d86eb38ade service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Received event network-vif-deleted-3fc25179-89cf-42de-b6b6-f31806bcbce3 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1166.491560] env[70013]: INFO nova.compute.manager [req-a0b6c14b-24f7-4288-bd18-9c4c685171cc req-e732f3f1-112e-4915-9bba-b0d86eb38ade service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Neutron deleted interface 3fc25179-89cf-42de-b6b6-f31806bcbce3; detaching it from the instance and deleting it from the info cache [ 1166.491773] env[70013]: DEBUG nova.network.neutron [req-a0b6c14b-24f7-4288-bd18-9c4c685171cc req-e732f3f1-112e-4915-9bba-b0d86eb38ade service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.505237] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1166.505729] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1166.505911] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1166.506111] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1166.506265] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1166.506414] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1166.506624] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1166.506783] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1166.506954] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1166.507556] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1166.507556] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1166.513849] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-013f3828-e7a3-4492-9530-75b386e582de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.530340] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1166.530340] env[70013]: value = "task-4231555" [ 1166.530340] env[70013]: _type = "Task" [ 1166.530340] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.539350] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231555, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.572342] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-28a56a75-9b30-4121-8252-a9e57287441c" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.572808] env[70013]: DEBUG nova.objects.instance [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'migration_context' on Instance uuid 28a56a75-9b30-4121-8252-a9e57287441c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.616270] env[70013]: DEBUG nova.network.neutron [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.638011] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fd979911-bc43-404c-adcb-e2ed10bacdbd tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "9026841c-7bda-41e5-a4ac-03d0d3e37560" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.613s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.768581] env[70013]: DEBUG oslo_vmware.api [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.862380] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4fe6b5-f34d-4d6d-9722-8d0317af179a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.870837] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-510db8bb-478a-456a-b520-6cb5a7672cbb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.903322] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cfa545b-c90e-4a27-96e5-236f430d6c37 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.911813] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df704554-fe03-4e90-b69b-b411f03d3bde {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.933603] env[70013]: DEBUG nova.compute.provider_tree [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1166.994211] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3eae48c-6449-460c-a834-8fa8e71fa911 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.007126] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58f7833-0b08-435b-8960-90c136e34811 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.046523] env[70013]: DEBUG nova.compute.manager [req-a0b6c14b-24f7-4288-bd18-9c4c685171cc req-e732f3f1-112e-4915-9bba-b0d86eb38ade service nova] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Detach interface failed, port_id=3fc25179-89cf-42de-b6b6-f31806bcbce3, reason: Instance 472b8218-51c3-492e-96ff-5fa99df4cc4c could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1167.054324] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231555, 'name': ReconfigVM_Task, 'duration_secs': 0.176794} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.055225] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance 'dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8' progress to 33 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1167.076055] env[70013]: DEBUG nova.objects.base [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Object Instance<28a56a75-9b30-4121-8252-a9e57287441c> lazy-loaded attributes: info_cache,migration_context {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1167.077035] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e14cdc-3ae9-4526-ab38-20024529b9d8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.102680] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2bb06fc-e1ac-449e-b10e-d0f3d61dd677 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.109302] env[70013]: DEBUG oslo_vmware.api [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1167.109302] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d63eeb-0fa5-2ba1-986e-31c8b3ca8653" [ 1167.109302] env[70013]: _type = "Task" [ 1167.109302] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.120253] env[70013]: INFO nova.compute.manager [-] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Took 1.47 seconds to deallocate network for instance. [ 1167.120861] env[70013]: DEBUG oslo_vmware.api [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d63eeb-0fa5-2ba1-986e-31c8b3ca8653, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.187300] env[70013]: DEBUG nova.network.neutron [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updated VIF entry in instance network info cache for port b209fd87-7748-481e-8262-5ec43c4a436b. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1167.187777] env[70013]: DEBUG nova.network.neutron [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b209fd87-7748-481e-8262-5ec43c4a436b", "address": "fa:16:3e:08:bd:72", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb209fd87-77", "ovs_interfaceid": "b209fd87-7748-481e-8262-5ec43c4a436b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.260865] env[70013]: DEBUG oslo_vmware.api [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231554, 'name': ReconfigVM_Task, 'duration_secs': 0.768409} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.261479] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.261743] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Reconfigured VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1167.437658] env[70013]: DEBUG nova.scheduler.client.report [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1167.563292] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1167.563558] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1167.563721] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1167.563904] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1167.564070] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1167.564221] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1167.564427] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1167.564582] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1167.564745] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1167.564908] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1167.565099] env[70013]: DEBUG nova.virt.hardware [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1167.570397] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1167.570703] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e0e9aa1-3a28-4d32-afe0-a90177177d20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.590224] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1167.590224] env[70013]: value = "task-4231556" [ 1167.590224] env[70013]: _type = "Task" [ 1167.590224] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.598481] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.620433] env[70013]: DEBUG oslo_vmware.api [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d63eeb-0fa5-2ba1-986e-31c8b3ca8653, 'name': SearchDatastore_Task, 'duration_secs': 0.009427} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.620796] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.626986] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.691023] env[70013]: DEBUG oslo_concurrency.lockutils [req-02f6408e-9481-4548-bb47-2a99f207290e req-c1d1de14-aec1-4a98-9687-239dc253eb05 service nova] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.766531] env[70013]: DEBUG oslo_concurrency.lockutils [None req-8662308a-4a45-43db-8717-c8c2eb32945c tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-38738d80-25ca-4a17-81f1-e0a7bc8c4a44-b209fd87-7748-481e-8262-5ec43c4a436b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.200s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.943534] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.946674] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.326s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.963669] env[70013]: INFO nova.scheduler.client.report [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Deleted allocations for instance 09544922-6dc4-48b5-8cfd-e91e7f74c13f [ 1168.101703] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231556, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.473928] env[70013]: DEBUG oslo_concurrency.lockutils [None req-fca506ee-be02-4f80-96f0-9cf9ffd3b144 tempest-ServersTestJSON-1431904245 tempest-ServersTestJSON-1431904245-project-member] Lock "09544922-6dc4-48b5-8cfd-e91e7f74c13f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.114s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.607274] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231556, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.676757] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82d6737-eedd-4ab2-b0ab-ab464da49719 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.684958] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60f9781-55f6-4bbe-ad1e-51cc66e274ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.715492] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddb0711-5d27-4d13-bcb8-b85a1b612e25 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.723768] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9bd613-6242-40c5-9526-93bda0197a3b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.738557] env[70013]: DEBUG nova.compute.provider_tree [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.102568] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231556, 'name': ReconfigVM_Task, 'duration_secs': 1.215241} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.102857] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1169.103674] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a6e811-b7e5-41b2-b8eb-3698f3214354 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.127849] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8/dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1169.128203] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-934a2630-7f3c-4459-a42b-33f9f11b4595 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.147112] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1169.147112] env[70013]: value = "task-4231557" [ 1169.147112] env[70013]: _type = "Task" [ 1169.147112] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.155594] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231557, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.242475] env[70013]: DEBUG nova.scheduler.client.report [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1169.296483] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-38738d80-25ca-4a17-81f1-e0a7bc8c4a44-b209fd87-7748-481e-8262-5ec43c4a436b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1169.296764] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-38738d80-25ca-4a17-81f1-e0a7bc8c4a44-b209fd87-7748-481e-8262-5ec43c4a436b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.658083] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231557, 'name': ReconfigVM_Task, 'duration_secs': 0.279029} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.658380] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Reconfigured VM instance instance-0000006a to attach disk [datastore2] dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8/dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1169.658653] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance 'dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8' progress to 50 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1169.799623] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.799941] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.800661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61516107-4fa0-4567-9c40-701bfb366485 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.820448] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07247d38-0757-45b7-845f-d0223bb23e75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.847887] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Reconfiguring VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1169.847981] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55c5d324-1c55-410f-a913-a4e08d435742 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.868627] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1169.868627] env[70013]: value = "task-4231558" [ 1169.868627] env[70013]: _type = "Task" [ 1169.868627] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.877595] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.166187] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8047ff-681f-44d7-a8ff-66ab35037ce0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.189659] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e62281-e31f-4ded-b28a-547a6ec5cac0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.208865] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance 'dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8' progress to 67 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1170.254094] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.307s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1170.257015] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.630s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1170.257279] env[70013]: DEBUG nova.objects.instance [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'resources' on Instance uuid 472b8218-51c3-492e-96ff-5fa99df4cc4c {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.379786] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.757420] env[70013]: DEBUG nova.network.neutron [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Port 43a074aa-e66b-4258-8c74-fae23dca9c71 binding to destination host cpu-1 is already ACTIVE {{(pid=70013) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1170.870568] env[70013]: INFO nova.scheduler.client.report [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted allocation for migration 8f1251c5-4dcb-4b6f-afa5-81f973aaec05 [ 1170.886240] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.027732] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6c17dc-a751-4334-9a39-e51f3964d902 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.036869] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a2afb3-ba14-41a7-8cac-f7e4a9e18e58 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.074632] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933a8799-c4f6-4d5a-96d8-b1a2c147aac2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.083878] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b614938-95b3-4172-a861-aa21b01af15b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.099057] env[70013]: DEBUG nova.compute.provider_tree [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1171.272310] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.382925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b11d2f94-bdf9-4d44-8cbd-2ef92fe2aebe tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.453s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.384054] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.384313] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.112s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.384525] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "28a56a75-9b30-4121-8252-a9e57287441c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.384731] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.384898] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.387919] env[70013]: INFO nova.compute.manager [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Terminating instance [ 1171.601843] env[70013]: DEBUG nova.scheduler.client.report [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1171.785965] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.786244] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.786430] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.883563] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.891904] env[70013]: DEBUG nova.compute.manager [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1171.892210] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1171.893226] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601bb91d-a5e3-4cfb-8b13-74412f1c7bdf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.901719] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1171.901986] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee4d8c87-e4e0-4994-861e-465b591d033d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.909035] env[70013]: DEBUG oslo_vmware.api [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1171.909035] env[70013]: value = "task-4231559" [ 1171.909035] env[70013]: _type = "Task" [ 1171.909035] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.919122] env[70013]: DEBUG oslo_vmware.api [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231559, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.107287] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.850s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.137362] env[70013]: INFO nova.scheduler.client.report [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted allocations for instance 472b8218-51c3-492e-96ff-5fa99df4cc4c [ 1172.385505] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.422435] env[70013]: DEBUG oslo_vmware.api [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231559, 'name': PowerOffVM_Task, 'duration_secs': 0.204782} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.422967] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1172.423265] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1172.423643] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd5a5723-9a00-4c26-b673-a4cdbb1f8ab0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.520063] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1172.520452] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1172.520707] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleting the datastore file [datastore2] 28a56a75-9b30-4121-8252-a9e57287441c {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1172.521032] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfb3acd2-4b5c-41b4-a20e-bb0e4a63eb59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.529806] env[70013]: DEBUG oslo_vmware.api [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1172.529806] env[70013]: value = "task-4231561" [ 1172.529806] env[70013]: _type = "Task" [ 1172.529806] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.540053] env[70013]: DEBUG oslo_vmware.api [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231561, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.645705] env[70013]: DEBUG oslo_concurrency.lockutils [None req-42879cfb-ceb2-4b7e-9e78-ae536d398e30 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "472b8218-51c3-492e-96ff-5fa99df4cc4c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.641s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.836929] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.837149] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1172.837349] env[70013]: DEBUG nova.network.neutron [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1172.885502] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.039833] env[70013]: DEBUG oslo_vmware.api [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231561, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165197} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.040112] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.040306] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.040484] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.040663] env[70013]: INFO nova.compute.manager [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1173.040919] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1173.041138] env[70013]: DEBUG nova.compute.manager [-] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1173.041237] env[70013]: DEBUG nova.network.neutron [-] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1173.386899] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.583678] env[70013]: DEBUG nova.network.neutron [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance_info_cache with network_info: [{"id": "43a074aa-e66b-4258-8c74-fae23dca9c71", "address": "fa:16:3e:2d:02:dc", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a074aa-e6", "ovs_interfaceid": "43a074aa-e66b-4258-8c74-fae23dca9c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.761432] env[70013]: DEBUG nova.compute.manager [req-e15b9cfb-3a62-47a0-8fa7-b5ec12b23337 req-6aa38585-f360-4d9f-acc9-ce739f834f92 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Received event network-vif-deleted-a0d11944-d177-42a5-ad2b-22c116396f8a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1173.761432] env[70013]: INFO nova.compute.manager [req-e15b9cfb-3a62-47a0-8fa7-b5ec12b23337 req-6aa38585-f360-4d9f-acc9-ce739f834f92 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Neutron deleted interface a0d11944-d177-42a5-ad2b-22c116396f8a; detaching it from the instance and deleting it from the info cache [ 1173.761432] env[70013]: DEBUG nova.network.neutron [req-e15b9cfb-3a62-47a0-8fa7-b5ec12b23337 req-6aa38585-f360-4d9f-acc9-ce739f834f92 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.888088] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.086475] env[70013]: DEBUG oslo_concurrency.lockutils [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1174.239216] env[70013]: DEBUG nova.network.neutron [-] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.265928] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b355ad21-26f1-4044-b826-7bdfe9ee3373 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.283226] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2600226-c25c-4c6c-8ab4-bd15ebc8d98a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.316499] env[70013]: DEBUG nova.compute.manager [req-e15b9cfb-3a62-47a0-8fa7-b5ec12b23337 req-6aa38585-f360-4d9f-acc9-ce739f834f92 service nova] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Detach interface failed, port_id=a0d11944-d177-42a5-ad2b-22c116396f8a, reason: Instance 28a56a75-9b30-4121-8252-a9e57287441c could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1174.387569] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.615737] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752bd78a-0a94-4ac9-9241-e344a50ed186 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.638949] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b168dd27-beed-4118-b2a2-4e374eecc683 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.646698] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance 'dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8' progress to 83 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1174.742473] env[70013]: INFO nova.compute.manager [-] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Took 1.70 seconds to deallocate network for instance. [ 1174.887085] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.153697] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1175.154144] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35357cf5-29cb-4af0-9b10-288e734ed1c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.161735] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1175.161735] env[70013]: value = "task-4231562" [ 1175.161735] env[70013]: _type = "Task" [ 1175.161735] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.170978] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231562, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.249550] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.249863] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.250070] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.268245] env[70013]: INFO nova.scheduler.client.report [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted allocations for instance 28a56a75-9b30-4121-8252-a9e57287441c [ 1175.301603] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.301837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.390157] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.672564] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231562, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.778813] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e66508c1-5380-482a-98e6-a2dadb035aed tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "28a56a75-9b30-4121-8252-a9e57287441c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.394s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.804502] env[70013]: DEBUG nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1175.888793] env[70013]: DEBUG oslo_vmware.api [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231558, 'name': ReconfigVM_Task, 'duration_secs': 5.786035} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.889066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1175.889295] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Reconfigured VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1176.172793] env[70013]: DEBUG oslo_vmware.api [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231562, 'name': PowerOnVM_Task, 'duration_secs': 0.893628} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.173211] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1176.173376] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-660e5445-e750-4c0d-985c-a0921e45c550 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance 'dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8' progress to 100 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1176.325527] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.325804] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.327501] env[70013]: INFO nova.compute.claims [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1176.349998] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.350262] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.852857] env[70013]: DEBUG nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1177.195107] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.195381] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1177.195522] env[70013]: DEBUG nova.network.neutron [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1177.370898] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.532414] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdeb1510-1c6d-4f1e-97cb-9c59d4d5f85f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.540969] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099f00e9-e112-4b40-b28c-8892b13eb9ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.572181] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4e8779-62ff-40aa-a23a-fb1f489128e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.581508] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34dd28ed-aaea-4ecb-921f-80f4bc92873d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.595842] env[70013]: DEBUG nova.compute.provider_tree [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.908848] env[70013]: INFO nova.network.neutron [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Port b209fd87-7748-481e-8262-5ec43c4a436b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1177.909236] env[70013]: DEBUG nova.network.neutron [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.018972] env[70013]: DEBUG nova.compute.manager [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1178.019193] env[70013]: DEBUG nova.compute.manager [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing instance network info cache due to event network-changed-4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1178.019740] env[70013]: DEBUG oslo_concurrency.lockutils [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] Acquiring lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.099601] env[70013]: DEBUG nova.scheduler.client.report [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1178.411925] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.414285] env[70013]: DEBUG oslo_concurrency.lockutils [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] Acquired lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.414476] env[70013]: DEBUG nova.network.neutron [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Refreshing network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1178.501862] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.502158] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.502352] env[70013]: DEBUG nova.compute.manager [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Going to confirm migration 3 {{(pid=70013) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1178.604872] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.605551] env[70013]: DEBUG nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1178.608415] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.238s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.610763] env[70013]: INFO nova.compute.claims [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1178.715579] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-1eea3d47-589d-450b-b5f8-2e763cd1bbd3-b209fd87-7748-481e-8262-5ec43c4a436b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.715856] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1eea3d47-589d-450b-b5f8-2e763cd1bbd3-b209fd87-7748-481e-8262-5ec43c4a436b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.716281] env[70013]: DEBUG nova.objects.instance [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'flavor' on Instance uuid 1eea3d47-589d-450b-b5f8-2e763cd1bbd3 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1178.916981] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d63aab9d-8487-43ed-83cd-c1deb8eb4d19 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-38738d80-25ca-4a17-81f1-e0a7bc8c4a44-b209fd87-7748-481e-8262-5ec43c4a436b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.620s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.065077] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.065163] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1179.065295] env[70013]: DEBUG nova.network.neutron [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1179.065478] env[70013]: DEBUG nova.objects.instance [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'info_cache' on Instance uuid dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1179.115283] env[70013]: DEBUG nova.compute.utils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1179.120027] env[70013]: DEBUG nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1179.120027] env[70013]: DEBUG nova.network.neutron [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1179.121406] env[70013]: DEBUG nova.network.neutron [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updated VIF entry in instance network info cache for port 4d8f0295-7d55-4b26-803e-5f289211ed13. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1179.121790] env[70013]: DEBUG nova.network.neutron [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [{"id": "4d8f0295-7d55-4b26-803e-5f289211ed13", "address": "fa:16:3e:a4:cb:14", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d8f0295-7d", "ovs_interfaceid": "4d8f0295-7d55-4b26-803e-5f289211ed13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.161376] env[70013]: DEBUG nova.policy [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1634fc5b4baa40c6bd9e3bef9eb77864', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f5b0bd88f5046f098579c5d59e3f0c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1179.312869] env[70013]: DEBUG nova.objects.instance [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'pci_requests' on Instance uuid 1eea3d47-589d-450b-b5f8-2e763cd1bbd3 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1179.619377] env[70013]: DEBUG nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1179.626754] env[70013]: DEBUG oslo_concurrency.lockutils [req-cc8aa7a6-2b5c-497f-b43a-e41f222f8bc7 req-8eee8553-ae94-40de-8b82-1a09fbc5d17f service nova] Releasing lock "refresh_cache-38738d80-25ca-4a17-81f1-e0a7bc8c4a44" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1179.672651] env[70013]: DEBUG nova.network.neutron [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Successfully created port: b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1179.817412] env[70013]: DEBUG nova.objects.base [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Object Instance<1eea3d47-589d-450b-b5f8-2e763cd1bbd3> lazy-loaded attributes: flavor,pci_requests {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1179.817634] env[70013]: DEBUG nova.network.neutron [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1179.843166] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7593429d-d502-4c12-b7d0-c92362488d5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.852765] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb159b9c-f962-4734-8b22-fad630d7b10e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.883955] env[70013]: DEBUG nova.policy [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b702df24aa64b2e97fb515f5e0891a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52386d3aa84b4d3e8376dec7544fae6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1179.886383] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4982bb-c599-4e73-97e3-8750c94d10d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.894567] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132eb79f-06df-47db-8d64-d2cf7bacb7ed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.908039] env[70013]: DEBUG nova.compute.provider_tree [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.291510] env[70013]: DEBUG nova.network.neutron [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance_info_cache with network_info: [{"id": "43a074aa-e66b-4258-8c74-fae23dca9c71", "address": "fa:16:3e:2d:02:dc", "network": {"id": "616efc9a-ffea-4886-91ab-eb073a21dd37", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-393103639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3d42e73dc0a4a06a2022d8b54e13e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a074aa-e6", "ovs_interfaceid": "43a074aa-e66b-4258-8c74-fae23dca9c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.411879] env[70013]: DEBUG nova.scheduler.client.report [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1180.630731] env[70013]: DEBUG nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1180.656969] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1180.657176] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1180.657337] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1180.657520] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1180.657664] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1180.657810] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1180.658023] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1180.658189] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1180.658356] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1180.658521] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1180.658691] env[70013]: DEBUG nova.virt.hardware [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1180.659559] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c35c8f-f41a-4638-9b6e-7de6a5c7a03b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.669740] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ed9633-7448-4ab3-a13b-f5efe134be9d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.794951] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1180.795268] env[70013]: DEBUG nova.objects.instance [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'migration_context' on Instance uuid dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1180.917612] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.918187] env[70013]: DEBUG nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1181.134719] env[70013]: DEBUG nova.network.neutron [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Successfully updated port: b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1181.298372] env[70013]: DEBUG nova.objects.base [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1181.299389] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8b7d46-254b-4bc4-a37d-5e8773c55e3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.320042] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae8bba0c-eb40-40b9-aa70-a9de66885544 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.328312] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1181.328312] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52235ea2-2b59-84d6-414a-a0420b25cca2" [ 1181.328312] env[70013]: _type = "Task" [ 1181.328312] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.336814] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52235ea2-2b59-84d6-414a-a0420b25cca2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.393182] env[70013]: DEBUG nova.network.neutron [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Successfully updated port: b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1181.423339] env[70013]: DEBUG nova.compute.utils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1181.424864] env[70013]: DEBUG nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1181.425079] env[70013]: DEBUG nova.network.neutron [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1181.462136] env[70013]: DEBUG nova.policy [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d6c00fc04b8435388ae5e2652f8e0eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51b2e9e21dcc447d86aa3d2f5f307d4f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1181.637413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.637783] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1181.637854] env[70013]: DEBUG nova.network.neutron [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1181.743054] env[70013]: DEBUG nova.network.neutron [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Successfully created port: 461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1181.839727] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52235ea2-2b59-84d6-414a-a0420b25cca2, 'name': SearchDatastore_Task, 'duration_secs': 0.010239} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.840050] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.840347] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.895548] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.895915] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1181.895915] env[70013]: DEBUG nova.network.neutron [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1181.928239] env[70013]: DEBUG nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1182.188212] env[70013]: DEBUG nova.network.neutron [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1182.332052] env[70013]: DEBUG nova.network.neutron [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": "b9586fe7-a244-4613-a095-504009e343a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.433635] env[70013]: WARNING nova.network.neutron [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] b160f70a-0663-495b-bd35-440c047157c8 already exists in list: networks containing: ['b160f70a-0663-495b-bd35-440c047157c8']. ignoring it [ 1182.581237] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da295219-7928-43b5-8982-8308eb344664 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.588830] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee21d3c5-0a2c-4f7d-92f6-01103db30c9f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.621888] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753a67da-64dc-4b54-8139-3a8f9840fe97 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.630137] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6d7641-d226-4a01-87a0-7c4cfe170743 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.644572] env[70013]: DEBUG nova.compute.provider_tree [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1182.771218] env[70013]: DEBUG nova.network.neutron [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b209fd87-7748-481e-8262-5ec43c4a436b", "address": "fa:16:3e:08:bd:72", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb209fd87-77", "ovs_interfaceid": "b209fd87-7748-481e-8262-5ec43c4a436b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.834539] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1182.834882] env[70013]: DEBUG nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Instance network_info: |[{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": "b9586fe7-a244-4613-a095-504009e343a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1182.835402] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:8a:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9586fe7-a244-4613-a095-504009e343a0', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1182.842722] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1182.842984] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1182.843228] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1833fd2-642e-4c33-966d-5ff2409b85ae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.864235] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1182.864235] env[70013]: value = "task-4231563" [ 1182.864235] env[70013]: _type = "Task" [ 1182.864235] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.875094] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231563, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.944318] env[70013]: DEBUG nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1182.973444] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1182.973753] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1182.974069] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1182.974339] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1182.974504] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1182.974675] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1182.974877] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1182.975110] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1182.975361] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1182.975566] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1182.975781] env[70013]: DEBUG nova.virt.hardware [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1182.976707] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d2f388-ab17-44b0-93e4-605fabfff7d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.987801] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d5e38c-1d3e-4f84-bb6f-34b2136614a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.150130] env[70013]: DEBUG nova.scheduler.client.report [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1183.259288] env[70013]: DEBUG nova.network.neutron [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Successfully updated port: 461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1183.274236] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.274944] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.275181] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.276341] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4e0436-a1fb-42c4-855f-f9099811e61e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.295750] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1183.296061] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1183.296207] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1183.296404] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1183.296550] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1183.296696] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1183.296900] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1183.297082] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1183.297326] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1183.297513] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1183.297692] env[70013]: DEBUG nova.virt.hardware [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1183.304201] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Reconfiguring VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1183.304588] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbc74dba-ee8c-4caf-836c-8312a0be9044 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.323287] env[70013]: DEBUG oslo_vmware.api [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1183.323287] env[70013]: value = "task-4231564" [ 1183.323287] env[70013]: _type = "Task" [ 1183.323287] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.332464] env[70013]: DEBUG oslo_vmware.api [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231564, 'name': ReconfigVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.375592] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231563, 'name': CreateVM_Task, 'duration_secs': 0.345214} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.375803] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1183.376552] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.376674] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.377148] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1183.377442] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-280f8043-7001-4b8f-8e93-0ed139ef6e15 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.382758] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1183.382758] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a9b9e6-88a0-594f-ac30-5a2ff0853082" [ 1183.382758] env[70013]: _type = "Task" [ 1183.382758] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.391628] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a9b9e6-88a0-594f-ac30-5a2ff0853082, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.762993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.763427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.763468] env[70013]: DEBUG nova.network.neutron [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1183.834140] env[70013]: DEBUG oslo_vmware.api [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231564, 'name': ReconfigVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.893751] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52a9b9e6-88a0-594f-ac30-5a2ff0853082, 'name': SearchDatastore_Task, 'duration_secs': 0.009151} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.894102] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.894350] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1183.894592] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.894739] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.894919] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1183.895207] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17d71cc5-0a00-4db7-bbad-2abc88bf7f4e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.904240] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1183.904476] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1183.905273] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-859ebfc9-f4de-4e58-88db-6b4d8504a737 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.911605] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1183.911605] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6dc62-051b-24d8-27cf-c6535c6ef582" [ 1183.911605] env[70013]: _type = "Task" [ 1183.911605] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.920401] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6dc62-051b-24d8-27cf-c6535c6ef582, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.161218] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.321s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.306442] env[70013]: DEBUG nova.network.neutron [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1184.333967] env[70013]: DEBUG oslo_vmware.api [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231564, 'name': ReconfigVM_Task, 'duration_secs': 0.564663} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.334603] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.334882] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Reconfigured VM to attach interface {{(pid=70013) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1184.425020] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6dc62-051b-24d8-27cf-c6535c6ef582, 'name': SearchDatastore_Task, 'duration_secs': 0.015439} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.425827] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f621f3a6-0676-40d7-a767-6a86ddd0e0b2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.431242] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1184.431242] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5283d98f-d416-2380-a2b8-2fdd6135b084" [ 1184.431242] env[70013]: _type = "Task" [ 1184.431242] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.439768] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5283d98f-d416-2380-a2b8-2fdd6135b084, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.450853] env[70013]: DEBUG nova.network.neutron [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.718722] env[70013]: INFO nova.scheduler.client.report [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted allocation for migration 2d43898c-e276-4439-863b-6de02d09ffa7 [ 1184.840492] env[70013]: DEBUG oslo_concurrency.lockutils [None req-74c1a98c-7049-4f1a-adc5-aba3d0bdd01a tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1eea3d47-589d-450b-b5f8-2e763cd1bbd3-b209fd87-7748-481e-8262-5ec43c4a436b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.124s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.941830] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5283d98f-d416-2380-a2b8-2fdd6135b084, 'name': SearchDatastore_Task, 'duration_secs': 0.010312} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.942055] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.942324] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1184.942583] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7509e84-a8d3-4795-bcf8-e23426bdd9bb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.950268] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1184.950268] env[70013]: value = "task-4231565" [ 1184.950268] env[70013]: _type = "Task" [ 1184.950268] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.954946] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.955304] env[70013]: DEBUG nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Instance network_info: |[{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1184.955756] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:00:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'abcf0d10-3f3f-45dc-923e-1c78766e2dad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '461fa18b-f3db-4db6-ba87-fe872af49659', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1184.963229] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1184.963857] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1184.964432] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca2ebf84-5aac-4f8e-b9ad-21ae0709b1a2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.981539] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231565, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.987623] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1184.987623] env[70013]: value = "task-4231566" [ 1184.987623] env[70013]: _type = "Task" [ 1184.987623] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.995876] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231566, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.225917] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.724s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1185.460988] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231565, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.497908] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231566, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.962042] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231565, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.000544] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231566, 'name': CreateVM_Task} progress is 25%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.055582] env[70013]: DEBUG nova.compute.manager [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1186.055918] env[70013]: DEBUG nova.compute.manager [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing instance network info cache due to event network-changed-dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1186.056284] env[70013]: DEBUG oslo_concurrency.lockutils [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.056347] env[70013]: DEBUG oslo_concurrency.lockutils [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1186.056977] env[70013]: DEBUG nova.network.neutron [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1186.075661] env[70013]: DEBUG nova.compute.manager [req-1e03846c-7b0d-4ce5-8f8f-7462403f9daf req-af561d6d-ce8e-49ff-92c3-7cd70148c769 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-vif-plugged-b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1186.076217] env[70013]: DEBUG oslo_concurrency.lockutils [req-1e03846c-7b0d-4ce5-8f8f-7462403f9daf req-af561d6d-ce8e-49ff-92c3-7cd70148c769 service nova] Acquiring lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.076217] env[70013]: DEBUG oslo_concurrency.lockutils [req-1e03846c-7b0d-4ce5-8f8f-7462403f9daf req-af561d6d-ce8e-49ff-92c3-7cd70148c769 service nova] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.076428] env[70013]: DEBUG oslo_concurrency.lockutils [req-1e03846c-7b0d-4ce5-8f8f-7462403f9daf req-af561d6d-ce8e-49ff-92c3-7cd70148c769 service nova] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.076578] env[70013]: DEBUG nova.compute.manager [req-1e03846c-7b0d-4ce5-8f8f-7462403f9daf req-af561d6d-ce8e-49ff-92c3-7cd70148c769 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] No waiting events found dispatching network-vif-plugged-b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1186.076748] env[70013]: WARNING nova.compute.manager [req-1e03846c-7b0d-4ce5-8f8f-7462403f9daf req-af561d6d-ce8e-49ff-92c3-7cd70148c769 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received unexpected event network-vif-plugged-b209fd87-7748-481e-8262-5ec43c4a436b for instance with vm_state active and task_state None. [ 1186.463055] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231565, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.125161} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.463333] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1186.463563] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1186.463817] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a678a711-2fa8-423a-9af9-80f5c3524a0b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.470853] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1186.470853] env[70013]: value = "task-4231567" [ 1186.470853] env[70013]: _type = "Task" [ 1186.470853] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.478716] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231567, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.499487] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231566, 'name': CreateVM_Task, 'duration_secs': 1.248528} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.499664] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1186.500388] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.500558] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1186.500984] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1186.501211] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a200ffd7-013d-49b8-b31d-d5076036df4f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.509912] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1186.509912] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a9ee6-cfd7-12a2-a66f-f7446f5ae602" [ 1186.509912] env[70013]: _type = "Task" [ 1186.509912] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.520920] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528a9ee6-cfd7-12a2-a66f-f7446f5ae602, 'name': SearchDatastore_Task, 'duration_secs': 0.010231} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.521794] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1186.522053] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1186.522294] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.522447] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1186.522631] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1186.523158] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0f892cc-158f-4b13-96cf-3d3734e9b197 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.526808] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.527019] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.527252] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.527461] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.527630] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.529608] env[70013]: INFO nova.compute.manager [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Terminating instance [ 1186.533176] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1186.533360] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1186.534252] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfaccebb-2af4-4377-b929-8e1abaf755ce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.540722] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1186.540722] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52093800-a16b-f9b8-0636-6bf96bd3ceba" [ 1186.540722] env[70013]: _type = "Task" [ 1186.540722] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.549776] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52093800-a16b-f9b8-0636-6bf96bd3ceba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.778185] env[70013]: DEBUG nova.network.neutron [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updated VIF entry in instance network info cache for port dcb5c6dd-7799-4712-91fd-86a53ac56170. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1186.778627] env[70013]: DEBUG nova.network.neutron [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b209fd87-7748-481e-8262-5ec43c4a436b", "address": "fa:16:3e:08:bd:72", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb209fd87-77", "ovs_interfaceid": "b209fd87-7748-481e-8262-5ec43c4a436b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.805278] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "interface-1eea3d47-589d-450b-b5f8-2e763cd1bbd3-b209fd87-7748-481e-8262-5ec43c4a436b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.805528] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1eea3d47-589d-450b-b5f8-2e763cd1bbd3-b209fd87-7748-481e-8262-5ec43c4a436b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.981072] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231567, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081933} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.981350] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1186.982166] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80566d1f-a3e7-491e-b884-a912429abe97 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.004236] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1187.004481] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69c70510-de0e-4fbe-8f60-6acc6c779f9f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.018844] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "499bb115-146c-48dc-9447-3c01c9549a73" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.019106] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "499bb115-146c-48dc-9447-3c01c9549a73" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.019320] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "499bb115-146c-48dc-9447-3c01c9549a73-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.019503] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "499bb115-146c-48dc-9447-3c01c9549a73-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.019666] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "499bb115-146c-48dc-9447-3c01c9549a73-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1187.021856] env[70013]: INFO nova.compute.manager [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Terminating instance [ 1187.027973] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1187.027973] env[70013]: value = "task-4231568" [ 1187.027973] env[70013]: _type = "Task" [ 1187.027973] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.037995] env[70013]: DEBUG nova.compute.manager [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1187.038321] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1187.038907] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231568, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.039684] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fd622c-29dc-425a-9b10-e2d6809aa76b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.051497] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52093800-a16b-f9b8-0636-6bf96bd3ceba, 'name': SearchDatastore_Task, 'duration_secs': 0.010389} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.054270] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.054792] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dae63b1-d230-42e0-8b01-ba27a927e50f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.057377] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3a8cfb3-7550-4de2-88f7-e6675b8304a8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.062418] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1187.062418] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d22d43-98c2-e336-960d-5e82fac05a20" [ 1187.062418] env[70013]: _type = "Task" [ 1187.062418] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.066889] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1187.066889] env[70013]: value = "task-4231569" [ 1187.066889] env[70013]: _type = "Task" [ 1187.066889] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.073149] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d22d43-98c2-e336-960d-5e82fac05a20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.078197] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.282064] env[70013]: DEBUG oslo_concurrency.lockutils [req-4aa82f82-b51b-42e4-a8d4-94ad9175d383 req-e48fe305-42a5-483d-9917-e604ebb94c5e service nova] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.308823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.309037] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1187.309957] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd859e63-9588-4ff4-9910-494f34f4d38c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.328672] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bc167e-04f6-47c5-a411-46c184304ee4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.355168] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Reconfiguring VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1187.355693] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdcdb159-9fc6-4ef8-a834-dc839efacce9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.377033] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1187.377033] env[70013]: value = "task-4231570" [ 1187.377033] env[70013]: _type = "Task" [ 1187.377033] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.385293] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.526472] env[70013]: DEBUG nova.compute.manager [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1187.526694] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1187.527628] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474f1a21-783e-41e5-b22a-b69039c61275 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.539465] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231568, 'name': ReconfigVM_Task, 'duration_secs': 0.290651} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.541567] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1187.542261] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.542484] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c7b22cb-145f-4449-8ff6-010f7e092c33 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.543973] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce17ef3e-36f9-499d-b18e-7a2039441212 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.551706] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1187.551706] env[70013]: value = "task-4231572" [ 1187.551706] env[70013]: _type = "Task" [ 1187.551706] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.552317] env[70013]: DEBUG oslo_vmware.api [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1187.552317] env[70013]: value = "task-4231571" [ 1187.552317] env[70013]: _type = "Task" [ 1187.552317] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.564927] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231572, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.571525] env[70013]: DEBUG oslo_vmware.api [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.580967] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d22d43-98c2-e336-960d-5e82fac05a20, 'name': SearchDatastore_Task, 'duration_secs': 0.010482} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.584330] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.584609] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1187.584925] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231569, 'name': PowerOffVM_Task, 'duration_secs': 0.204986} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.585162] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d94ec7ef-2f7f-4bd4-bc48-f3c61d208cf3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.587493] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1187.587686] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.587929] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d6e8088-88e3-4004-a87a-facf5fcfdf9d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.595326] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1187.595326] env[70013]: value = "task-4231573" [ 1187.595326] env[70013]: _type = "Task" [ 1187.595326] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.604260] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.657941] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.658281] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.658522] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleting the datastore file [datastore2] dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.658849] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27b12a02-6d98-4974-9866-305dd0b44f62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.667153] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for the task: (returnval){ [ 1187.667153] env[70013]: value = "task-4231575" [ 1187.667153] env[70013]: _type = "Task" [ 1187.667153] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.677628] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.888297] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.067740] env[70013]: DEBUG oslo_vmware.api [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231571, 'name': PowerOffVM_Task, 'duration_secs': 0.195825} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.073331] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1188.073688] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1188.074596] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231572, 'name': Rename_Task, 'duration_secs': 0.161365} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.074887] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e38cda8-5247-42bf-b7ac-29432034c8f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.077486] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1188.077693] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5988de7-32cb-4c9e-813c-42cf658f82ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.086420] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1188.086420] env[70013]: value = "task-4231576" [ 1188.086420] env[70013]: _type = "Task" [ 1188.086420] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.088846] env[70013]: DEBUG nova.compute.manager [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-vif-plugged-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1188.089089] env[70013]: DEBUG oslo_concurrency.lockutils [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.089305] env[70013]: DEBUG oslo_concurrency.lockutils [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.089567] env[70013]: DEBUG oslo_concurrency.lockutils [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.089668] env[70013]: DEBUG nova.compute.manager [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] No waiting events found dispatching network-vif-plugged-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1188.089803] env[70013]: WARNING nova.compute.manager [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received unexpected event network-vif-plugged-b9586fe7-a244-4613-a095-504009e343a0 for instance with vm_state building and task_state spawning. [ 1188.089963] env[70013]: DEBUG nova.compute.manager [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-changed-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1188.090137] env[70013]: DEBUG nova.compute.manager [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing instance network info cache due to event network-changed-b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1188.090327] env[70013]: DEBUG oslo_concurrency.lockutils [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] Acquiring lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.090465] env[70013]: DEBUG oslo_concurrency.lockutils [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] Acquired lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1188.090620] env[70013]: DEBUG nova.network.neutron [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing network info cache for port b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1188.102854] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231576, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.109655] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231573, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499263} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.109929] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1188.110165] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1188.110420] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e4916d4-8a33-45dc-b847-dd5e54575e1d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.115114] env[70013]: DEBUG nova.compute.manager [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-changed-b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1188.115332] env[70013]: DEBUG nova.compute.manager [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing instance network info cache due to event network-changed-b209fd87-7748-481e-8262-5ec43c4a436b. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1188.115577] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.115724] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1188.115885] env[70013]: DEBUG nova.network.neutron [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Refreshing network info cache for port b209fd87-7748-481e-8262-5ec43c4a436b {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1188.123970] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1188.123970] env[70013]: value = "task-4231578" [ 1188.123970] env[70013]: _type = "Task" [ 1188.123970] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.135159] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231578, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.159878] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1188.160162] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1188.160336] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Deleting the datastore file [datastore2] 499bb115-146c-48dc-9447-3c01c9549a73 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1188.160981] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2317860-96ed-459b-9543-03349754c17a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.168216] env[70013]: DEBUG oslo_vmware.api [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for the task: (returnval){ [ 1188.168216] env[70013]: value = "task-4231579" [ 1188.168216] env[70013]: _type = "Task" [ 1188.168216] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.180212] env[70013]: DEBUG oslo_vmware.api [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Task: {'id': task-4231575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.417805} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.183310] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.183433] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.183622] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.183806] env[70013]: INFO nova.compute.manager [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1188.184059] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1188.184272] env[70013]: DEBUG oslo_vmware.api [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.184484] env[70013]: DEBUG nova.compute.manager [-] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1188.184586] env[70013]: DEBUG nova.network.neutron [-] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1188.389340] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.599537] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231576, 'name': PowerOnVM_Task} progress is 33%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.634161] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231578, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068662} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.636743] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1188.637579] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b435d049-eb6c-4205-87e7-a90e07d0bdc2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.660985] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.663701] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae438cb9-8b38-40e1-a5f5-373147a2f9ad {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.686640] env[70013]: DEBUG oslo_vmware.api [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Task: {'id': task-4231579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176142} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.688144] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.688260] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.688977] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.688977] env[70013]: INFO nova.compute.manager [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1188.688977] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1188.689157] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1188.689157] env[70013]: value = "task-4231580" [ 1188.689157] env[70013]: _type = "Task" [ 1188.689157] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.691565] env[70013]: DEBUG nova.compute.manager [-] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1188.691669] env[70013]: DEBUG nova.network.neutron [-] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1188.702298] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231580, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.891316] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.917576] env[70013]: DEBUG nova.network.neutron [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updated VIF entry in instance network info cache for port b209fd87-7748-481e-8262-5ec43c4a436b. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1188.917898] env[70013]: DEBUG nova.network.neutron [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b209fd87-7748-481e-8262-5ec43c4a436b", "address": "fa:16:3e:08:bd:72", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb209fd87-77", "ovs_interfaceid": "b209fd87-7748-481e-8262-5ec43c4a436b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.926089] env[70013]: DEBUG nova.network.neutron [-] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.931137] env[70013]: DEBUG nova.network.neutron [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updated VIF entry in instance network info cache for port b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1188.931137] env[70013]: DEBUG nova.network.neutron [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": "b9586fe7-a244-4613-a095-504009e343a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.100940] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231576, 'name': PowerOnVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.209222] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231580, 'name': ReconfigVM_Task, 'duration_secs': 0.468039} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.209515] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1189.213143] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae5a33fc-46f5-4ecc-89f6-115938850789 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.218076] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1189.218076] env[70013]: value = "task-4231581" [ 1189.218076] env[70013]: _type = "Task" [ 1189.218076] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.228228] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231581, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.390128] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.424737] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1189.425051] env[70013]: DEBUG nova.compute.manager [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Received event network-vif-plugged-461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1189.425302] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Acquiring lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.425532] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.425864] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.425864] env[70013]: DEBUG nova.compute.manager [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] No waiting events found dispatching network-vif-plugged-461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1189.426038] env[70013]: WARNING nova.compute.manager [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Received unexpected event network-vif-plugged-461fa18b-f3db-4db6-ba87-fe872af49659 for instance with vm_state building and task_state spawning. [ 1189.426203] env[70013]: DEBUG nova.compute.manager [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Received event network-changed-461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1189.426361] env[70013]: DEBUG nova.compute.manager [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Refreshing instance network info cache due to event network-changed-461fa18b-f3db-4db6-ba87-fe872af49659. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1189.426551] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Acquiring lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.426691] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Acquired lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.426847] env[70013]: DEBUG nova.network.neutron [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Refreshing network info cache for port 461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1189.429737] env[70013]: INFO nova.compute.manager [-] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Took 1.25 seconds to deallocate network for instance. [ 1189.436790] env[70013]: DEBUG oslo_concurrency.lockutils [req-e4e7ec79-ae96-4306-84a6-78829e20e1a8 req-0873dac6-22dd-4cf4-9024-c957f070fb8a service nova] Releasing lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1189.598548] env[70013]: DEBUG oslo_vmware.api [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231576, 'name': PowerOnVM_Task, 'duration_secs': 1.242734} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.598822] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1189.599039] env[70013]: INFO nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Took 8.97 seconds to spawn the instance on the hypervisor. [ 1189.599235] env[70013]: DEBUG nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1189.599997] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf5181f-b8e3-477b-874e-614f05561513 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.728464] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231581, 'name': Rename_Task, 'duration_secs': 0.259264} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.728761] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1189.729033] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c4fde23-b660-4d41-a142-20b7a227333d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.737262] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1189.737262] env[70013]: value = "task-4231582" [ 1189.737262] env[70013]: _type = "Task" [ 1189.737262] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.748902] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.774092] env[70013]: DEBUG nova.network.neutron [-] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.894823] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.940352] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.940656] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.940938] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.964826] env[70013]: INFO nova.scheduler.client.report [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted allocations for instance dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8 [ 1190.119188] env[70013]: INFO nova.compute.manager [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Took 13.81 seconds to build instance. [ 1190.144289] env[70013]: DEBUG nova.compute.manager [req-4af832e6-ca33-411a-b9e9-aff0937ff081 req-4d6dfe4b-e819-47c2-83c0-358d56e0baa7 service nova] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Received event network-vif-deleted-43a074aa-e66b-4258-8c74-fae23dca9c71 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1190.144500] env[70013]: DEBUG nova.compute.manager [req-4af832e6-ca33-411a-b9e9-aff0937ff081 req-4d6dfe4b-e819-47c2-83c0-358d56e0baa7 service nova] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Received event network-vif-deleted-cc3af4a9-2d59-4b89-975c-87e4a5e6d939 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1190.153298] env[70013]: DEBUG nova.network.neutron [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updated VIF entry in instance network info cache for port 461fa18b-f3db-4db6-ba87-fe872af49659. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1190.153639] env[70013]: DEBUG nova.network.neutron [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.250741] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231582, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.277200] env[70013]: INFO nova.compute.manager [-] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Took 1.59 seconds to deallocate network for instance. [ 1190.392032] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.475343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f09a4564-1b2a-42a3-be30-822db010f127 tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.948s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.621928] env[70013]: DEBUG oslo_concurrency.lockutils [None req-990d2cfc-4c29-4eb7-901b-0b6bcd63dfc6 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.320s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.657034] env[70013]: DEBUG oslo_concurrency.lockutils [req-d9c7286b-e0aa-441f-893a-b20976cd3b58 req-9c0cc219-64c7-4379-aafa-a2082ae57ec1 service nova] Releasing lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1190.749741] env[70013]: DEBUG oslo_vmware.api [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231582, 'name': PowerOnVM_Task, 'duration_secs': 0.557389} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.750143] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1190.750362] env[70013]: INFO nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Took 7.81 seconds to spawn the instance on the hypervisor. [ 1190.750549] env[70013]: DEBUG nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1190.751412] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451095f1-0de4-4c73-95dc-39893ae2d507 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.784080] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.784434] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1190.784667] env[70013]: DEBUG nova.objects.instance [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lazy-loading 'resources' on Instance uuid 499bb115-146c-48dc-9447-3c01c9549a73 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1190.892626] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.054471] env[70013]: DEBUG nova.compute.manager [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-changed-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1191.054663] env[70013]: DEBUG nova.compute.manager [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing instance network info cache due to event network-changed-b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1191.054936] env[70013]: DEBUG oslo_concurrency.lockutils [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] Acquiring lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.055227] env[70013]: DEBUG oslo_concurrency.lockutils [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] Acquired lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1191.055370] env[70013]: DEBUG nova.network.neutron [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing network info cache for port b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1191.272700] env[70013]: INFO nova.compute.manager [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Took 13.91 seconds to build instance. [ 1191.317541] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "fc473d40-b57b-437e-9511-58a0ba700a69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.318271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.318657] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.320099] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.320099] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.322179] env[70013]: INFO nova.compute.manager [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Terminating instance [ 1191.392494] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.470709] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b295740-2a15-4b30-9c68-d5abaa8adfea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.479393] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee6d4eb-8f4a-4d0a-baa8-55f9300c7ef0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.513502] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f05b49-16e8-4e3f-9df9-8bf0d14ab22f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.522404] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fce2c11-95fe-4c42-9f44-15baa8af535e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.538656] env[70013]: DEBUG nova.compute.provider_tree [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.774870] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4fd4fead-b073-457d-9d7d-e6f3ccc19916 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.424s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.781270] env[70013]: DEBUG nova.network.neutron [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updated VIF entry in instance network info cache for port b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1191.781665] env[70013]: DEBUG nova.network.neutron [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": "b9586fe7-a244-4613-a095-504009e343a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.829081] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.829295] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquired lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1191.829482] env[70013]: DEBUG nova.network.neutron [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1191.894193] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.041668] env[70013]: DEBUG nova.scheduler.client.report [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1192.284708] env[70013]: DEBUG oslo_concurrency.lockutils [req-52859784-2699-43f4-8d5a-b3260b5cac14 req-350fe171-8293-4161-a7f3-2fa0985537ea service nova] Releasing lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.355758] env[70013]: DEBUG nova.network.neutron [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1192.393795] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.480480] env[70013]: DEBUG nova.network.neutron [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.549578] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1192.587937] env[70013]: INFO nova.scheduler.client.report [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Deleted allocations for instance 499bb115-146c-48dc-9447-3c01c9549a73 [ 1192.896710] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task} progress is 18%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.985144] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Releasing lock "refresh_cache-fc473d40-b57b-437e-9511-58a0ba700a69" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.985683] env[70013]: DEBUG nova.compute.manager [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1192.985916] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1192.986298] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c27dfe48-e643-4629-a67a-bd1bb37c5b78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.998781] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219a652f-cda9-4829-b46d-cadab9eb22d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.028427] env[70013]: WARNING nova.virt.vmwareapi.vmops [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fc473d40-b57b-437e-9511-58a0ba700a69 could not be found. [ 1193.028771] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1193.029052] env[70013]: INFO nova.compute.manager [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1193.029418] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1193.029760] env[70013]: DEBUG nova.compute.manager [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1193.029916] env[70013]: DEBUG nova.network.neutron [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1193.052562] env[70013]: DEBUG nova.network.neutron [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1193.088985] env[70013]: DEBUG nova.compute.manager [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Received event network-changed-461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1193.089991] env[70013]: DEBUG nova.compute.manager [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Refreshing instance network info cache due to event network-changed-461fa18b-f3db-4db6-ba87-fe872af49659. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1193.089991] env[70013]: DEBUG oslo_concurrency.lockutils [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] Acquiring lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.089991] env[70013]: DEBUG oslo_concurrency.lockutils [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] Acquired lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1193.089991] env[70013]: DEBUG nova.network.neutron [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Refreshing network info cache for port 461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1193.096086] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b5460b06-ca45-4802-8160-2a8298f53c95 tempest-ServersTestJSON-294753548 tempest-ServersTestJSON-294753548-project-member] Lock "499bb115-146c-48dc-9447-3c01c9549a73" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.077s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.396395] env[70013]: DEBUG oslo_vmware.api [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231570, 'name': ReconfigVM_Task, 'duration_secs': 5.79303} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.397194] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.397681] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Reconfigured VM to detach interface {{(pid=70013) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1193.556150] env[70013]: DEBUG nova.network.neutron [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.934747] env[70013]: DEBUG nova.network.neutron [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updated VIF entry in instance network info cache for port 461fa18b-f3db-4db6-ba87-fe872af49659. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1193.935068] env[70013]: DEBUG nova.network.neutron [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1194.059491] env[70013]: INFO nova.compute.manager [-] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Took 1.03 seconds to deallocate network for instance. [ 1194.438164] env[70013]: DEBUG oslo_concurrency.lockutils [req-ee8f4b54-2268-4209-ae96-d9a264871b24 req-f48ed8a9-d448-4497-bb93-00e9f5787c32 service nova] Releasing lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1194.570664] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1194.571024] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1194.571267] env[70013]: DEBUG nova.objects.instance [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lazy-loading 'resources' on Instance uuid fc473d40-b57b-437e-9511-58a0ba700a69 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.898758] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.898758] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquired lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1194.898758] env[70013]: DEBUG nova.network.neutron [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1195.338364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.338364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.338364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.338364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.338364] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.339230] env[70013]: INFO nova.compute.manager [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Terminating instance [ 1195.356020] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea01d4bc-6c44-4db8-81d0-21d91245477a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.363143] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca6bfc5-f886-4959-a8ae-08300436f17a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.395160] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d91407-c6e1-4154-8537-b9d6d21e6fc1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.405477] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2cf765a-2af9-4e99-a6aa-be93809f6602 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.421125] env[70013]: DEBUG nova.compute.provider_tree [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1195.612537] env[70013]: INFO nova.network.neutron [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Port b209fd87-7748-481e-8262-5ec43c4a436b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1195.613919] env[70013]: DEBUG nova.network.neutron [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [{"id": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "address": "fa:16:3e:59:9a:0a", "network": {"id": "b160f70a-0663-495b-bd35-440c047157c8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-809441840-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52386d3aa84b4d3e8376dec7544fae6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcb5c6dd-77", "ovs_interfaceid": "dcb5c6dd-7799-4712-91fd-86a53ac56170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.843080] env[70013]: DEBUG nova.compute.manager [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1195.843345] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1195.844317] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95eb48a4-e1af-4642-aeb2-fe55a6fe1834 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.852831] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1195.853146] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-319949de-ba62-42f1-8723-ae509c360e8d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.860440] env[70013]: DEBUG oslo_vmware.api [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1195.860440] env[70013]: value = "task-4231583" [ 1195.860440] env[70013]: _type = "Task" [ 1195.860440] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.872159] env[70013]: DEBUG oslo_vmware.api [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.925198] env[70013]: DEBUG nova.scheduler.client.report [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1196.118419] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Releasing lock "refresh_cache-1eea3d47-589d-450b-b5f8-2e763cd1bbd3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.375408] env[70013]: DEBUG oslo_vmware.api [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231583, 'name': PowerOffVM_Task, 'duration_secs': 0.265896} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.375704] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.375872] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.376237] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb35c993-966f-4e07-ba1d-12532f4d8911 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.431172] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1196.443302] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.443763] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.443898] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleting the datastore file [datastore2] 1eea3d47-589d-450b-b5f8-2e763cd1bbd3 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.444091] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-753b3821-199a-443c-9a06-89931483e32d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.451112] env[70013]: DEBUG oslo_vmware.api [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1196.451112] env[70013]: value = "task-4231585" [ 1196.451112] env[70013]: _type = "Task" [ 1196.451112] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.461903] env[70013]: DEBUG oslo_vmware.api [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.462975] env[70013]: INFO nova.scheduler.client.report [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Deleted allocations for instance fc473d40-b57b-437e-9511-58a0ba700a69 [ 1196.623896] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0469f9d-a25c-40c9-9894-6c842f5d4869 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "interface-1eea3d47-589d-450b-b5f8-2e763cd1bbd3-b209fd87-7748-481e-8262-5ec43c4a436b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.818s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1196.962241] env[70013]: DEBUG oslo_vmware.api [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174782} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.962594] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1196.962808] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1196.963087] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1196.963261] env[70013]: INFO nova.compute.manager [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1196.963516] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1196.963720] env[70013]: DEBUG nova.compute.manager [-] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1196.963816] env[70013]: DEBUG nova.network.neutron [-] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1196.970817] env[70013]: DEBUG oslo_concurrency.lockutils [None req-80617bfa-7ac5-4115-b4cb-023c1d4fd3fb tempest-DeleteServersTestJSON-472565240 tempest-DeleteServersTestJSON-472565240-project-member] Lock "fc473d40-b57b-437e-9511-58a0ba700a69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.652s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.665059] env[70013]: DEBUG nova.compute.manager [req-252e2de6-40d0-476a-8efc-70e7c0a086f4 req-66d4cb40-68db-4a60-a654-f66702fafea8 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Received event network-vif-deleted-dcb5c6dd-7799-4712-91fd-86a53ac56170 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1197.665334] env[70013]: INFO nova.compute.manager [req-252e2de6-40d0-476a-8efc-70e7c0a086f4 req-66d4cb40-68db-4a60-a654-f66702fafea8 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Neutron deleted interface dcb5c6dd-7799-4712-91fd-86a53ac56170; detaching it from the instance and deleting it from the info cache [ 1197.665531] env[70013]: DEBUG nova.network.neutron [req-252e2de6-40d0-476a-8efc-70e7c0a086f4 req-66d4cb40-68db-4a60-a654-f66702fafea8 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.709685] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "60c2d10b-2fba-4341-8679-136d79fa8d53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1197.709919] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.148274] env[70013]: DEBUG nova.network.neutron [-] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.168327] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8f0fe95-033b-417c-83fa-7bdc546a2497 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.179302] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29149217-d823-4294-ad04-35c4401e99cd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.209933] env[70013]: DEBUG nova.compute.manager [req-252e2de6-40d0-476a-8efc-70e7c0a086f4 req-66d4cb40-68db-4a60-a654-f66702fafea8 service nova] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Detach interface failed, port_id=dcb5c6dd-7799-4712-91fd-86a53ac56170, reason: Instance 1eea3d47-589d-450b-b5f8-2e763cd1bbd3 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1198.212715] env[70013]: DEBUG nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1198.654249] env[70013]: INFO nova.compute.manager [-] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Took 1.69 seconds to deallocate network for instance. [ 1198.739637] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.740031] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.741631] env[70013]: INFO nova.compute.claims [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1199.161815] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.920684] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d140b2-4641-4ebe-abaf-d8efd497a0aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.929014] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57fd90c-c568-496d-93e3-82adb2c6269a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.959638] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b82215-dcd2-4ea7-ad64-03b60675fe51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.967778] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b84762-3a5d-419d-a4e8-7919042c4bdf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.983149] env[70013]: DEBUG nova.compute.provider_tree [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.486907] env[70013]: DEBUG nova.scheduler.client.report [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1200.993074] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.993534] env[70013]: DEBUG nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1200.996298] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.835s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.996575] env[70013]: DEBUG nova.objects.instance [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'resources' on Instance uuid 1eea3d47-589d-450b-b5f8-2e763cd1bbd3 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.500085] env[70013]: DEBUG nova.compute.utils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1201.501533] env[70013]: DEBUG nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1201.501733] env[70013]: DEBUG nova.network.neutron [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1201.548287] env[70013]: DEBUG nova.policy [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a28ad9b870ef4d20890b68adb9ca0ba6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f93fa571ef3453f9b035b5309dd67d8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1201.668859] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb94466-486b-4e04-b8ee-a39485af77c9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.677023] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e297277-0c28-473c-9168-c721e0666e3f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.708107] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf63c93-53ad-40a1-9c1c-374a19c939d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.715961] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c9f3e5-c926-42a7-939c-986f22dfa62a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.732785] env[70013]: DEBUG nova.compute.provider_tree [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.816933] env[70013]: DEBUG nova.network.neutron [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Successfully created port: a6ca2770-f8f4-4f87-9224-4a013733c952 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1202.007529] env[70013]: DEBUG nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1202.237035] env[70013]: DEBUG nova.scheduler.client.report [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1202.742677] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.746s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.763527] env[70013]: INFO nova.scheduler.client.report [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted allocations for instance 1eea3d47-589d-450b-b5f8-2e763cd1bbd3 [ 1203.017951] env[70013]: DEBUG nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1203.045952] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1203.046244] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1203.046388] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1203.046574] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1203.046721] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1203.046953] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1203.047258] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1203.047480] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1203.047711] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1203.047939] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1203.048204] env[70013]: DEBUG nova.virt.hardware [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1203.049314] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02017e21-f508-45e8-a5d6-03c29cabaffa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.058883] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8774a7ab-c4f1-40f7-9ad4-7099d821b253 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.204909] env[70013]: DEBUG nova.compute.manager [req-6f7cdf62-4fbf-45a3-9422-372e384fd926 req-a780edf8-f239-4bfa-a091-ac56a1d896a8 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Received event network-vif-plugged-a6ca2770-f8f4-4f87-9224-4a013733c952 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1203.205150] env[70013]: DEBUG oslo_concurrency.lockutils [req-6f7cdf62-4fbf-45a3-9422-372e384fd926 req-a780edf8-f239-4bfa-a091-ac56a1d896a8 service nova] Acquiring lock "60c2d10b-2fba-4341-8679-136d79fa8d53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1203.205399] env[70013]: DEBUG oslo_concurrency.lockutils [req-6f7cdf62-4fbf-45a3-9422-372e384fd926 req-a780edf8-f239-4bfa-a091-ac56a1d896a8 service nova] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.205522] env[70013]: DEBUG oslo_concurrency.lockutils [req-6f7cdf62-4fbf-45a3-9422-372e384fd926 req-a780edf8-f239-4bfa-a091-ac56a1d896a8 service nova] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.205696] env[70013]: DEBUG nova.compute.manager [req-6f7cdf62-4fbf-45a3-9422-372e384fd926 req-a780edf8-f239-4bfa-a091-ac56a1d896a8 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] No waiting events found dispatching network-vif-plugged-a6ca2770-f8f4-4f87-9224-4a013733c952 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1203.205890] env[70013]: WARNING nova.compute.manager [req-6f7cdf62-4fbf-45a3-9422-372e384fd926 req-a780edf8-f239-4bfa-a091-ac56a1d896a8 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Received unexpected event network-vif-plugged-a6ca2770-f8f4-4f87-9224-4a013733c952 for instance with vm_state building and task_state spawning. [ 1203.271439] env[70013]: DEBUG oslo_concurrency.lockutils [None req-86b0ae9b-e43c-4dc0-aa35-483346a852f3 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "1eea3d47-589d-450b-b5f8-2e763cd1bbd3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.935s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.300669] env[70013]: DEBUG nova.network.neutron [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Successfully updated port: a6ca2770-f8f4-4f87-9224-4a013733c952 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1203.807458] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "refresh_cache-60c2d10b-2fba-4341-8679-136d79fa8d53" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.807571] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquired lock "refresh_cache-60c2d10b-2fba-4341-8679-136d79fa8d53" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1203.807765] env[70013]: DEBUG nova.network.neutron [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1204.248666] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1204.249067] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.249180] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1204.249324] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.249498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.251933] env[70013]: INFO nova.compute.manager [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Terminating instance [ 1204.338491] env[70013]: DEBUG nova.network.neutron [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1204.469137] env[70013]: DEBUG nova.network.neutron [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Updating instance_info_cache with network_info: [{"id": "a6ca2770-f8f4-4f87-9224-4a013733c952", "address": "fa:16:3e:2b:26:d3", "network": {"id": "456707e9-8287-4e4a-b3ea-b3ab867b1f81", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-529979210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f93fa571ef3453f9b035b5309dd67d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6ca2770-f8", "ovs_interfaceid": "a6ca2770-f8f4-4f87-9224-4a013733c952", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1204.755822] env[70013]: DEBUG nova.compute.manager [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1204.756062] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1204.756924] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbccdd8-b2d9-4909-80f7-5d4b2d9c30db {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.765279] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1204.765499] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e258b37b-b9ac-4a2c-afa4-671edb8c111a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.772551] env[70013]: DEBUG oslo_vmware.api [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1204.772551] env[70013]: value = "task-4231586" [ 1204.772551] env[70013]: _type = "Task" [ 1204.772551] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.781974] env[70013]: DEBUG oslo_vmware.api [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231586, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.971957] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Releasing lock "refresh_cache-60c2d10b-2fba-4341-8679-136d79fa8d53" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1204.972730] env[70013]: DEBUG nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Instance network_info: |[{"id": "a6ca2770-f8f4-4f87-9224-4a013733c952", "address": "fa:16:3e:2b:26:d3", "network": {"id": "456707e9-8287-4e4a-b3ea-b3ab867b1f81", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-529979210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f93fa571ef3453f9b035b5309dd67d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6ca2770-f8", "ovs_interfaceid": "a6ca2770-f8f4-4f87-9224-4a013733c952", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1204.973251] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:26:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53ebf5df-5ecb-4a0c-a163-d88165639de0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6ca2770-f8f4-4f87-9224-4a013733c952', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1204.980760] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Creating folder: Project (7f93fa571ef3453f9b035b5309dd67d8). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1204.981048] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2d85b08-020c-4029-9f78-2b0cec949639 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.993694] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Created folder: Project (7f93fa571ef3453f9b035b5309dd67d8) in parent group-v836999. [ 1204.993902] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Creating folder: Instances. Parent ref: group-v837233. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1204.994175] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c1226d3-4bc2-49e8-916a-9a16dd62884d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.006470] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Created folder: Instances in parent group-v837233. [ 1205.006926] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1205.007040] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1205.007298] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfaaadd7-a84b-471e-813a-e36393a977c2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.027604] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1205.027604] env[70013]: value = "task-4231589" [ 1205.027604] env[70013]: _type = "Task" [ 1205.027604] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.035454] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231589, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.232514] env[70013]: DEBUG nova.compute.manager [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Received event network-changed-a6ca2770-f8f4-4f87-9224-4a013733c952 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1205.232602] env[70013]: DEBUG nova.compute.manager [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Refreshing instance network info cache due to event network-changed-a6ca2770-f8f4-4f87-9224-4a013733c952. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1205.232850] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] Acquiring lock "refresh_cache-60c2d10b-2fba-4341-8679-136d79fa8d53" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.232954] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] Acquired lock "refresh_cache-60c2d10b-2fba-4341-8679-136d79fa8d53" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1205.233156] env[70013]: DEBUG nova.network.neutron [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Refreshing network info cache for port a6ca2770-f8f4-4f87-9224-4a013733c952 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1205.283324] env[70013]: DEBUG oslo_vmware.api [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231586, 'name': PowerOffVM_Task, 'duration_secs': 0.184248} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.283689] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1205.283732] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1205.283988] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f0b1d01-ce42-495d-9ce6-dbf472ab15df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.538135] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231589, 'name': CreateVM_Task, 'duration_secs': 0.359478} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.538201] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1205.538989] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.539225] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1205.539628] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1205.539928] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-279cdf67-9ac2-4dc0-a20b-5151f88461ff {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.544765] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1205.544765] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5236201e-3ec6-f6e0-cc4a-ee2321cb5b78" [ 1205.544765] env[70013]: _type = "Task" [ 1205.544765] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.553247] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5236201e-3ec6-f6e0-cc4a-ee2321cb5b78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.923219] env[70013]: DEBUG nova.network.neutron [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Updated VIF entry in instance network info cache for port a6ca2770-f8f4-4f87-9224-4a013733c952. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1205.923636] env[70013]: DEBUG nova.network.neutron [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Updating instance_info_cache with network_info: [{"id": "a6ca2770-f8f4-4f87-9224-4a013733c952", "address": "fa:16:3e:2b:26:d3", "network": {"id": "456707e9-8287-4e4a-b3ea-b3ab867b1f81", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-529979210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f93fa571ef3453f9b035b5309dd67d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6ca2770-f8", "ovs_interfaceid": "a6ca2770-f8f4-4f87-9224-4a013733c952", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.055867] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5236201e-3ec6-f6e0-cc4a-ee2321cb5b78, 'name': SearchDatastore_Task, 'duration_secs': 0.010279} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.056208] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1206.056452] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1206.056687] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.056891] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1206.057078] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1206.057298] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-450479f2-92ee-4342-aba1-3930907d39cd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.066398] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1206.066591] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1206.067316] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97fc5742-6dc1-4b7a-b7c9-837f9728c161 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.074497] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1206.074497] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d546b2-36bc-33ac-2a78-e78f16434933" [ 1206.074497] env[70013]: _type = "Task" [ 1206.074497] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.083092] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d546b2-36bc-33ac-2a78-e78f16434933, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.427869] env[70013]: DEBUG oslo_concurrency.lockutils [req-5f1ffe3c-fe7a-4f38-87cd-fbf2194b6678 req-af6ca46f-12d7-48aa-91af-372d0fadc7a2 service nova] Releasing lock "refresh_cache-60c2d10b-2fba-4341-8679-136d79fa8d53" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1206.460643] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1206.460866] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1206.461256] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleting the datastore file [datastore1] 38738d80-25ca-4a17-81f1-e0a7bc8c4a44 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1206.461639] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4416729a-e072-4be9-ab1b-b91bf7c77f84 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.468497] env[70013]: DEBUG oslo_vmware.api [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for the task: (returnval){ [ 1206.468497] env[70013]: value = "task-4231591" [ 1206.468497] env[70013]: _type = "Task" [ 1206.468497] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.477076] env[70013]: DEBUG oslo_vmware.api [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.586470] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d546b2-36bc-33ac-2a78-e78f16434933, 'name': SearchDatastore_Task, 'duration_secs': 0.009247} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.587094] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fb10dc1-6b4b-47ac-adcc-36c93b1caa45 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.593495] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1206.593495] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52252c21-b409-7762-a1da-68cbcf4ce4dd" [ 1206.593495] env[70013]: _type = "Task" [ 1206.593495] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.602323] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52252c21-b409-7762-a1da-68cbcf4ce4dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.978362] env[70013]: DEBUG oslo_vmware.api [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Task: {'id': task-4231591, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160786} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.978625] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.978808] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1206.978991] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1206.979188] env[70013]: INFO nova.compute.manager [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1206.979421] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1206.979617] env[70013]: DEBUG nova.compute.manager [-] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1206.979712] env[70013]: DEBUG nova.network.neutron [-] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1207.104123] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52252c21-b409-7762-a1da-68cbcf4ce4dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009514} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.104493] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1207.104774] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 60c2d10b-2fba-4341-8679-136d79fa8d53/60c2d10b-2fba-4341-8679-136d79fa8d53.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1207.105053] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4820e958-8a11-4880-8915-52def78bbcbc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.112541] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1207.112541] env[70013]: value = "task-4231592" [ 1207.112541] env[70013]: _type = "Task" [ 1207.112541] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.121792] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231592, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.260320] env[70013]: DEBUG nova.compute.manager [req-c321d99c-bcf6-4e9b-bc72-64a2c7bb89b8 req-833cd78a-42de-4270-b3ba-b8cf09904ee5 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Received event network-vif-deleted-4d8f0295-7d55-4b26-803e-5f289211ed13 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1207.260625] env[70013]: INFO nova.compute.manager [req-c321d99c-bcf6-4e9b-bc72-64a2c7bb89b8 req-833cd78a-42de-4270-b3ba-b8cf09904ee5 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Neutron deleted interface 4d8f0295-7d55-4b26-803e-5f289211ed13; detaching it from the instance and deleting it from the info cache [ 1207.260902] env[70013]: DEBUG nova.network.neutron [req-c321d99c-bcf6-4e9b-bc72-64a2c7bb89b8 req-833cd78a-42de-4270-b3ba-b8cf09904ee5 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.623230] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231592, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467716} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.623643] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 60c2d10b-2fba-4341-8679-136d79fa8d53/60c2d10b-2fba-4341-8679-136d79fa8d53.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1207.623737] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1207.624105] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f874428-02af-4571-9b84-97240f5d9dbd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.630873] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1207.630873] env[70013]: value = "task-4231593" [ 1207.630873] env[70013]: _type = "Task" [ 1207.630873] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.641109] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231593, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.705118] env[70013]: DEBUG nova.network.neutron [-] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.764555] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ec52c60-f788-4f48-a4f0-9448eb41c653 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.775258] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ae82ec-582e-49a3-828d-3f54234b71a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.804090] env[70013]: DEBUG nova.compute.manager [req-c321d99c-bcf6-4e9b-bc72-64a2c7bb89b8 req-833cd78a-42de-4270-b3ba-b8cf09904ee5 service nova] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Detach interface failed, port_id=4d8f0295-7d55-4b26-803e-5f289211ed13, reason: Instance 38738d80-25ca-4a17-81f1-e0a7bc8c4a44 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1208.141437] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231593, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064837} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.141709] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1208.142504] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da55d9e0-ca95-4c29-b3f6-9d9773f551ea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.164789] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 60c2d10b-2fba-4341-8679-136d79fa8d53/60c2d10b-2fba-4341-8679-136d79fa8d53.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.165110] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ee5b4c4-fe7b-4681-90c2-cee8c20263b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.185967] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1208.185967] env[70013]: value = "task-4231594" [ 1208.185967] env[70013]: _type = "Task" [ 1208.185967] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.194321] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.208307] env[70013]: INFO nova.compute.manager [-] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Took 1.23 seconds to deallocate network for instance. [ 1208.696316] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231594, 'name': ReconfigVM_Task, 'duration_secs': 0.292656} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.696666] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 60c2d10b-2fba-4341-8679-136d79fa8d53/60c2d10b-2fba-4341-8679-136d79fa8d53.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1208.697201] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e94f0c26-0da2-4494-90ab-87b91d76a7e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.704197] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1208.704197] env[70013]: value = "task-4231595" [ 1208.704197] env[70013]: _type = "Task" [ 1208.704197] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.716259] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1208.716522] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1208.716751] env[70013]: DEBUG nova.objects.instance [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lazy-loading 'resources' on Instance uuid 38738d80-25ca-4a17-81f1-e0a7bc8c4a44 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.717798] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231595, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.214307] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231595, 'name': Rename_Task, 'duration_secs': 0.146193} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.214639] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1209.214896] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-260cc64c-71a3-4d2a-b3a3-d25ecf849285 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.224056] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1209.224056] env[70013]: value = "task-4231596" [ 1209.224056] env[70013]: _type = "Task" [ 1209.224056] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.232832] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.378454] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021343c4-22cc-4190-9afe-5aaf5905c850 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.386366] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dec31fa-8a5b-4c6e-8d77-94b68612730c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.419125] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349fc308-fb9a-4154-a358-e55d26e49a18 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.427726] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4794a9db-2015-47ce-954b-b1df9981cc7b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.443286] env[70013]: DEBUG nova.compute.provider_tree [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.734093] env[70013]: DEBUG oslo_vmware.api [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231596, 'name': PowerOnVM_Task, 'duration_secs': 0.458551} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.734507] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1209.734613] env[70013]: INFO nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Took 6.72 seconds to spawn the instance on the hypervisor. [ 1209.734775] env[70013]: DEBUG nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1209.735565] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83692693-e79c-41f5-9871-43cc13cded4c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.947195] env[70013]: DEBUG nova.scheduler.client.report [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1210.253984] env[70013]: INFO nova.compute.manager [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Took 11.54 seconds to build instance. [ 1210.452305] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.736s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.475416] env[70013]: INFO nova.scheduler.client.report [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Deleted allocations for instance 38738d80-25ca-4a17-81f1-e0a7bc8c4a44 [ 1210.756152] env[70013]: DEBUG oslo_concurrency.lockutils [None req-90721d99-ccae-4766-8843-a7a94cd9c40e tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.046s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.844629] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "60c2d10b-2fba-4341-8679-136d79fa8d53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.844903] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.845144] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "60c2d10b-2fba-4341-8679-136d79fa8d53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.845341] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.845516] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.847629] env[70013]: INFO nova.compute.manager [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Terminating instance [ 1210.983585] env[70013]: DEBUG oslo_concurrency.lockutils [None req-02d19845-53fd-4098-ace6-ac37e48ef384 tempest-AttachInterfacesTestJSON-1305632387 tempest-AttachInterfacesTestJSON-1305632387-project-member] Lock "38738d80-25ca-4a17-81f1-e0a7bc8c4a44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.734s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.351466] env[70013]: DEBUG nova.compute.manager [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1211.351731] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1211.352723] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2f3a40-e7d9-4f69-934a-e141e8213f1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.362630] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1211.362888] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c2a5e0a-aedf-44ab-985b-2f8c1d68b52d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.369568] env[70013]: DEBUG oslo_vmware.api [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1211.369568] env[70013]: value = "task-4231597" [ 1211.369568] env[70013]: _type = "Task" [ 1211.369568] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.378959] env[70013]: DEBUG oslo_vmware.api [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231597, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.880344] env[70013]: DEBUG oslo_vmware.api [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231597, 'name': PowerOffVM_Task, 'duration_secs': 0.184548} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.880683] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1211.880780] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1211.881042] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-881d124d-fe18-494f-8131-7e3224d6faa0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.947196] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1211.947430] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1211.947621] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Deleting the datastore file [datastore2] 60c2d10b-2fba-4341-8679-136d79fa8d53 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1211.947897] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b954f70-eb37-4de3-b764-05a12a057789 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.957135] env[70013]: DEBUG oslo_vmware.api [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for the task: (returnval){ [ 1211.957135] env[70013]: value = "task-4231599" [ 1211.957135] env[70013]: _type = "Task" [ 1211.957135] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.965825] env[70013]: DEBUG oslo_vmware.api [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231599, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.165256] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.165511] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.165753] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.165905] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.166071] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.166223] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.166378] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.166512] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1212.166654] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.470024] env[70013]: DEBUG oslo_vmware.api [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Task: {'id': task-4231599, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143331} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.470024] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1212.470024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1212.470024] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1212.470024] env[70013]: INFO nova.compute.manager [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1212.470024] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1212.470024] env[70013]: DEBUG nova.compute.manager [-] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1212.470024] env[70013]: DEBUG nova.network.neutron [-] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1212.670243] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1212.670534] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1212.670718] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1212.670883] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1212.671819] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57c25e3-2126-44ad-a6bb-02787421f276 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.691350] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d3c2ce-fb29-491d-8e0f-452f1f51d226 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.718115] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505b5f29-6f7e-4fac-9e70-55bf32affc9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.733052] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26db59f-3c0f-4dcb-90a6-a397a3b474cf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.738412] env[70013]: DEBUG nova.compute.manager [req-c18c01f8-3a18-4dd3-9060-cfc031f1fa67 req-76d5a767-5ad5-42ed-b203-791f63f89c9e service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Received event network-vif-deleted-a6ca2770-f8f4-4f87-9224-4a013733c952 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1212.738646] env[70013]: INFO nova.compute.manager [req-c18c01f8-3a18-4dd3-9060-cfc031f1fa67 req-76d5a767-5ad5-42ed-b203-791f63f89c9e service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Neutron deleted interface a6ca2770-f8f4-4f87-9224-4a013733c952; detaching it from the instance and deleting it from the info cache [ 1212.738819] env[70013]: DEBUG nova.network.neutron [req-c18c01f8-3a18-4dd3-9060-cfc031f1fa67 req-76d5a767-5ad5-42ed-b203-791f63f89c9e service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.775874] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179330MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1212.776036] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1212.776230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1213.226706] env[70013]: DEBUG nova.network.neutron [-] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.243687] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb1c8891-a7e6-41ce-93df-510bd75aa80a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.252965] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13233b37-d55b-468e-9476-0d0e243b6cd1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.286443] env[70013]: DEBUG nova.compute.manager [req-c18c01f8-3a18-4dd3-9060-cfc031f1fa67 req-76d5a767-5ad5-42ed-b203-791f63f89c9e service nova] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Detach interface failed, port_id=a6ca2770-f8f4-4f87-9224-4a013733c952, reason: Instance 60c2d10b-2fba-4341-8679-136d79fa8d53 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1213.728388] env[70013]: INFO nova.compute.manager [-] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Took 1.26 seconds to deallocate network for instance. [ 1213.815735] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816045] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816045] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816183] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816280] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816483] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816651] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816651] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816735] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816841] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 351f5ef4-ed63-4d5d-b0a1-c029393898ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.816948] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 60c2d10b-2fba-4341-8679-136d79fa8d53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1213.817157] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1213.817299] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1213.971957] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e435d15d-57ee-4c1a-979f-e030c6ecd59c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.980222] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63950c30-e18b-4ec2-be3b-442ce431445e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.013548] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f318c8-5d11-49f6-9d9c-6f724e0b6396 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.022117] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a9ad01-a478-4e28-8f6f-263feda8ad62 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.037256] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1214.235902] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1214.540974] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1215.048039] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1215.048039] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.271s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1215.048039] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.812s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.048039] env[70013]: DEBUG nova.objects.instance [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lazy-loading 'resources' on Instance uuid 60c2d10b-2fba-4341-8679-136d79fa8d53 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.699548] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7247d6d1-61fd-45ac-aae4-6c0e184068a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.708382] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff9e6f8-6d56-4ebb-a4a9-78ab4c1483ee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.738909] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec2636d-a18f-474b-a0ec-3ce3306e8549 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.747224] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dbdd61-381f-48d3-8d8f-5cb8a3436814 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.762326] env[70013]: DEBUG nova.compute.provider_tree [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.265943] env[70013]: DEBUG nova.scheduler.client.report [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1216.770296] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1216.792189] env[70013]: INFO nova.scheduler.client.report [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Deleted allocations for instance 60c2d10b-2fba-4341-8679-136d79fa8d53 [ 1217.299294] env[70013]: DEBUG oslo_concurrency.lockutils [None req-46294b17-dd64-4e4d-ac93-34102c0f9a4a tempest-InstanceActionsNegativeTestJSON-99035032 tempest-InstanceActionsNegativeTestJSON-99035032-project-member] Lock "60c2d10b-2fba-4341-8679-136d79fa8d53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.454s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.024624] env[70013]: DEBUG nova.compute.manager [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Stashing vm_state: active {{(pid=70013) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1230.502397] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1230.502850] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.503119] env[70013]: INFO nova.compute.manager [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Shelving [ 1230.546356] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1230.546662] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1231.051784] env[70013]: INFO nova.compute.claims [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1231.513662] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1231.513995] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f7c8a17-d249-4d71-9fd0-16beb6dfb336 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.521681] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1231.521681] env[70013]: value = "task-4231600" [ 1231.521681] env[70013]: _type = "Task" [ 1231.521681] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.529728] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.559863] env[70013]: INFO nova.compute.resource_tracker [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating resource usage from migration 40a44e67-2ade-43ac-b4ca-2074cef220fd [ 1231.722411] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7984e15b-6882-46ad-b035-21f073e4cfa9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.730670] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4dcc5ad-d15b-45f7-b6df-16d98c1339de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.760915] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a22769c-fdbd-4423-adbe-a27c305f12de {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.769256] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f642e9-b8f7-4ca6-a593-caec359a9879 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.783038] env[70013]: DEBUG nova.compute.provider_tree [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1232.032081] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231600, 'name': PowerOffVM_Task, 'duration_secs': 0.191403} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.032443] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1232.033385] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cc305c-e907-4e34-bc80-a5f62b0bb2bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.054260] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e53bcd9-3b02-466b-8465-01f4df71bd13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.286228] env[70013]: DEBUG nova.scheduler.client.report [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1232.564910] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Creating Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1232.565210] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e24f688d-f7f6-4abc-a4bf-8cca02e0a2e9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.574333] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1232.574333] env[70013]: value = "task-4231601" [ 1232.574333] env[70013]: _type = "Task" [ 1232.574333] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.583222] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231601, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.791747] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.245s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1232.791998] env[70013]: INFO nova.compute.manager [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Migrating [ 1233.084598] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231601, 'name': CreateSnapshot_Task, 'duration_secs': 0.413552} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.085067] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Created Snapshot of the VM instance {{(pid=70013) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1233.085623] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914448f1-e847-4bf7-9752-06d2aba06150 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.307007] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.307007] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1233.307188] env[70013]: DEBUG nova.network.neutron [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1233.604080] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Creating linked-clone VM from snapshot {{(pid=70013) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1233.604448] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-83fdc0c6-938d-486f-958c-3cb75e494fd4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.614238] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1233.614238] env[70013]: value = "task-4231602" [ 1233.614238] env[70013]: _type = "Task" [ 1233.614238] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.623872] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231602, 'name': CloneVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.059325] env[70013]: DEBUG nova.network.neutron [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1234.126141] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231602, 'name': CloneVM_Task} progress is 94%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.562141] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1234.627542] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231602, 'name': CloneVM_Task, 'duration_secs': 0.983453} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.627813] env[70013]: INFO nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Created linked-clone VM from snapshot [ 1234.628600] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9899d30-2e87-4dae-bfd3-93c4679b594c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.636424] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Uploading image 4da16937-60b6-4c9f-b8ec-afe9cdd73248 {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1234.661191] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1234.661191] env[70013]: value = "vm-837237" [ 1234.661191] env[70013]: _type = "VirtualMachine" [ 1234.661191] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1234.661479] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-82df798e-4f80-4101-a6f0-9bb75b5ecb43 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.669749] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease: (returnval){ [ 1234.669749] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5293aad4-e790-59e3-553d-a40a78a11eee" [ 1234.669749] env[70013]: _type = "HttpNfcLease" [ 1234.669749] env[70013]: } obtained for exporting VM: (result){ [ 1234.669749] env[70013]: value = "vm-837237" [ 1234.669749] env[70013]: _type = "VirtualMachine" [ 1234.669749] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1234.670151] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the lease: (returnval){ [ 1234.670151] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5293aad4-e790-59e3-553d-a40a78a11eee" [ 1234.670151] env[70013]: _type = "HttpNfcLease" [ 1234.670151] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1234.677141] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1234.677141] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5293aad4-e790-59e3-553d-a40a78a11eee" [ 1234.677141] env[70013]: _type = "HttpNfcLease" [ 1234.677141] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1235.179080] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1235.179080] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5293aad4-e790-59e3-553d-a40a78a11eee" [ 1235.179080] env[70013]: _type = "HttpNfcLease" [ 1235.179080] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1235.179080] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1235.179080] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5293aad4-e790-59e3-553d-a40a78a11eee" [ 1235.179080] env[70013]: _type = "HttpNfcLease" [ 1235.179080] env[70013]: }. {{(pid=70013) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1235.179693] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d4361c-a3ea-4722-b101-be575b27c4ea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.188022] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524f9d39-874d-6eb1-1ae8-7e0fb2e3b5c8/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1235.188206] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524f9d39-874d-6eb1-1ae8-7e0fb2e3b5c8/disk-0.vmdk for reading. {{(pid=70013) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1235.273114] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1c8d8ecc-1d09-4312-924b-9a73621019ae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.080312] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d48d2f-c71b-48d4-b098-f68117127d30 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.101331] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance '351f5ef4-ed63-4d5d-b0a1-c029393898ac' progress to 0 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1236.608634] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1236.609051] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31939427-b17d-4d47-866f-7fc7b88ca6bd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.617554] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1236.617554] env[70013]: value = "task-4231604" [ 1236.617554] env[70013]: _type = "Task" [ 1236.617554] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.627582] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231604, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.128755] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231604, 'name': PowerOffVM_Task, 'duration_secs': 0.282092} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.129362] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1237.129727] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance '351f5ef4-ed63-4d5d-b0a1-c029393898ac' progress to 17 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1237.637429] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1237.637889] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1237.637889] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1237.638100] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1237.638255] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1237.638406] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1237.638608] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1237.638770] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1237.638937] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1237.639171] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1237.639420] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1237.644999] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d482598-e653-42a3-b794-a155caff8ca0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.663200] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1237.663200] env[70013]: value = "task-4231605" [ 1237.663200] env[70013]: _type = "Task" [ 1237.663200] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.675176] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.172947] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231605, 'name': ReconfigVM_Task, 'duration_secs': 0.483771} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.173409] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance '351f5ef4-ed63-4d5d-b0a1-c029393898ac' progress to 33 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1238.680925] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1238.681360] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1238.681400] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1238.681634] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1238.681811] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1238.681972] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1238.682198] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1238.682359] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1238.682523] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1238.682684] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1238.682858] env[70013]: DEBUG nova.virt.hardware [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1238.688362] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1238.688686] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c03b97ca-88e4-4119-8d0b-4ebd8d7879b9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.708657] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1238.708657] env[70013]: value = "task-4231606" [ 1238.708657] env[70013]: _type = "Task" [ 1238.708657] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.717842] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231606, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.219796] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231606, 'name': ReconfigVM_Task, 'duration_secs': 0.215799} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.220130] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1239.220995] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4892211a-c14a-462c-bcb7-5823d0d2d496 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.246070] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1239.246070] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1afa40c-f4d2-4351-a745-9abb0ef70e4f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.267975] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1239.267975] env[70013]: value = "task-4231607" [ 1239.267975] env[70013]: _type = "Task" [ 1239.267975] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.276206] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231607, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.778503] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231607, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.280524] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231607, 'name': ReconfigVM_Task, 'duration_secs': 1.007506} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.280848] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1240.281218] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance '351f5ef4-ed63-4d5d-b0a1-c029393898ac' progress to 50 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1240.339671] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.339856] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Cleaning up deleted instances with incomplete migration {{(pid=70013) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11902}} [ 1240.788860] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2be81c-3bd2-42ea-9981-46fb04c89af7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.808693] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebb19bc-4d59-422b-9313-0dda4461e7c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.828494] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance '351f5ef4-ed63-4d5d-b0a1-c029393898ac' progress to 67 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1240.841902] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.368589] env[70013]: DEBUG nova.network.neutron [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Port 461fa18b-f3db-4db6-ba87-fe872af49659 binding to destination host cpu-1 is already ACTIVE {{(pid=70013) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1242.344679] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.345103] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.345197] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.345285] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.345416] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1242.345572] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.391604] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.391883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.391993] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.848934] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.849231] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.849409] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.849571] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1242.850589] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e75beb-1d7b-4375-b4dc-1a575e781767 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.859696] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28efe77a-8b29-4dee-87da-11d5cea2b6d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.874710] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40085f9c-8c22-4584-9af4-c192f2629972 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.882239] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc29b35-5bc6-46f5-ba7a-2bb0f6fdf3cd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.914695] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179931MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1242.914853] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.915056] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.945741] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524f9d39-874d-6eb1-1ae8-7e0fb2e3b5c8/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1242.946763] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e677b2-88aa-44ba-b6b2-ac86626f7285 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.953798] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524f9d39-874d-6eb1-1ae8-7e0fb2e3b5c8/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1242.953979] env[70013]: ERROR oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524f9d39-874d-6eb1-1ae8-7e0fb2e3b5c8/disk-0.vmdk due to incomplete transfer. [ 1242.954220] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0105e4aa-9d71-4985-8249-bf1eaad4ee66 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.964694] env[70013]: DEBUG oslo_vmware.rw_handles [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524f9d39-874d-6eb1-1ae8-7e0fb2e3b5c8/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1242.964891] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Uploaded image 4da16937-60b6-4c9f-b8ec-afe9cdd73248 to the Glance image server {{(pid=70013) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1242.967231] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Destroying the VM {{(pid=70013) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1242.967693] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-271805e0-6a95-4242-88f7-47df90ab4625 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.974313] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1242.974313] env[70013]: value = "task-4231608" [ 1242.974313] env[70013]: _type = "Task" [ 1242.974313] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.983868] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231608, 'name': Destroy_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.453215] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.453505] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.453581] env[70013]: DEBUG nova.network.neutron [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1243.484845] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231608, 'name': Destroy_Task, 'duration_secs': 0.338123} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.485155] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Destroyed the VM [ 1243.485393] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Deleting Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1243.485647] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d72f656e-4457-4510-a504-3aca5493a2f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.492117] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1243.492117] env[70013]: value = "task-4231609" [ 1243.492117] env[70013]: _type = "Task" [ 1243.492117] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.500556] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231609, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.926810] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Applying migration context for instance 351f5ef4-ed63-4d5d-b0a1-c029393898ac as it has an incoming, in-progress migration 40a44e67-2ade-43ac-b4ca-2074cef220fd. Migration status is post-migrating {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1243.928195] env[70013]: INFO nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating resource usage from migration 40a44e67-2ade-43ac-b4ca-2074cef220fd [ 1243.949275] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.949413] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.949539] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.949657] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.949774] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.949885] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.949996] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.950139] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.950268] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.950404] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Migration 40a44e67-2ade-43ac-b4ca-2074cef220fd is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1742}} [ 1243.950519] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 351f5ef4-ed63-4d5d-b0a1-c029393898ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1243.950727] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1243.950864] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2752MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1244.004658] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231609, 'name': RemoveSnapshot_Task, 'duration_secs': 0.330559} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.008231] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Deleted Snapshot of the VM instance {{(pid=70013) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1244.008619] env[70013]: DEBUG nova.compute.manager [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1244.009946] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8944d99c-979f-4c50-a7d9-29b002ec0478 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.105124] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cf3bda-c982-48eb-bb8b-2298e6775d02 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.113228] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c280ab3b-eb88-4054-b10a-08f93db410dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.147409] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641eafe0-a256-401b-96c0-0a862230855d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.155564] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67bc2a1-2908-40f7-ace0-c25fadedf258 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.169752] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1244.235429] env[70013]: DEBUG nova.network.neutron [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1244.528024] env[70013]: INFO nova.compute.manager [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Shelve offloading [ 1244.672971] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1244.738169] env[70013]: DEBUG oslo_concurrency.lockutils [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1245.031382] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1245.031680] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7f26a2b-f54c-4b16-82dd-d3f9d8dde53f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.040156] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1245.040156] env[70013]: value = "task-4231610" [ 1245.040156] env[70013]: _type = "Task" [ 1245.040156] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.049625] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1245.049887] env[70013]: DEBUG nova.compute.manager [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1245.050661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fffb90-2922-42d4-8ab2-ba225e87334e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.056233] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1245.056397] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1245.056564] env[70013]: DEBUG nova.network.neutron [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1245.178067] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1245.178322] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.263s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1245.259424] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566ee681-cf27-438d-98be-ddfc4b704229 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.279086] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6acb9e-c30a-46f9-bf0b-c73db3561d94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.286352] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance '351f5ef4-ed63-4d5d-b0a1-c029393898ac' progress to 83 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1245.754770] env[70013]: DEBUG nova.network.neutron [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": "b9586fe7-a244-4613-a095-504009e343a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.792990] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1245.793316] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4842049-fb29-4296-a2f0-dc2b35bbfec6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.801992] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1245.801992] env[70013]: value = "task-4231611" [ 1245.801992] env[70013]: _type = "Task" [ 1245.801992] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.810371] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.168478] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.168788] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.258055] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1246.312076] env[70013]: DEBUG oslo_vmware.api [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231611, 'name': PowerOnVM_Task, 'duration_secs': 0.368829} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.312353] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1246.312571] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-45dc2232-859a-4f5d-bfea-c6d9bde52c34 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance '351f5ef4-ed63-4d5d-b0a1-c029393898ac' progress to 100 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1246.477820] env[70013]: DEBUG nova.compute.manager [req-1cb8e4b4-c52e-4cb2-b98b-4a10bb3d449b req-c3f61cd1-62c6-48ca-be71-c32d3c0f02a7 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-vif-unplugged-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1246.478105] env[70013]: DEBUG oslo_concurrency.lockutils [req-1cb8e4b4-c52e-4cb2-b98b-4a10bb3d449b req-c3f61cd1-62c6-48ca-be71-c32d3c0f02a7 service nova] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.478310] env[70013]: DEBUG oslo_concurrency.lockutils [req-1cb8e4b4-c52e-4cb2-b98b-4a10bb3d449b req-c3f61cd1-62c6-48ca-be71-c32d3c0f02a7 service nova] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.478483] env[70013]: DEBUG oslo_concurrency.lockutils [req-1cb8e4b4-c52e-4cb2-b98b-4a10bb3d449b req-c3f61cd1-62c6-48ca-be71-c32d3c0f02a7 service nova] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1246.478654] env[70013]: DEBUG nova.compute.manager [req-1cb8e4b4-c52e-4cb2-b98b-4a10bb3d449b req-c3f61cd1-62c6-48ca-be71-c32d3c0f02a7 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] No waiting events found dispatching network-vif-unplugged-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1246.478823] env[70013]: WARNING nova.compute.manager [req-1cb8e4b4-c52e-4cb2-b98b-4a10bb3d449b req-c3f61cd1-62c6-48ca-be71-c32d3c0f02a7 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received unexpected event network-vif-unplugged-b9586fe7-a244-4613-a095-504009e343a0 for instance with vm_state shelved and task_state shelving_offloading. [ 1246.567566] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1246.568590] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990d75eb-64b2-42a6-81ba-74e7e33365f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.576740] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1246.577012] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8d57270-401b-4d55-acca-60dda450dd67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.642861] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1246.643046] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1246.644046] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleting the datastore file [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1246.644046] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d694e9f5-69d2-49b0-9d6e-5c88f989d210 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.649486] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1246.649486] env[70013]: value = "task-4231613" [ 1246.649486] env[70013]: _type = "Task" [ 1246.649486] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.657738] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231613, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.676721] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.160083] env[70013]: DEBUG oslo_vmware.api [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231613, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133203} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.160083] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1247.160496] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1247.160496] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1247.260838] env[70013]: INFO nova.scheduler.client.report [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted allocations for instance 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 [ 1247.339545] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.339753] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Cleaning up deleted instances {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11864}} [ 1247.765301] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1247.765670] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1247.765979] env[70013]: DEBUG nova.objects.instance [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'resources' on Instance uuid 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.852698] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] There are 42 instances to clean {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11873}} [ 1247.852902] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 60c2d10b-2fba-4341-8679-136d79fa8d53] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1248.268786] env[70013]: DEBUG nova.objects.instance [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'numa_topology' on Instance uuid 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1248.357857] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: dbfe4e4b-85fd-49f2-a40d-0725a61eb1c8] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1248.480540] env[70013]: DEBUG nova.network.neutron [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Port 461fa18b-f3db-4db6-ba87-fe872af49659 binding to destination host cpu-1 is already ACTIVE {{(pid=70013) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1248.480860] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.481077] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1248.481282] env[70013]: DEBUG nova.network.neutron [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1248.504096] env[70013]: DEBUG nova.compute.manager [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-changed-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1248.504420] env[70013]: DEBUG nova.compute.manager [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing instance network info cache due to event network-changed-b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1248.504596] env[70013]: DEBUG oslo_concurrency.lockutils [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] Acquiring lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.504849] env[70013]: DEBUG oslo_concurrency.lockutils [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] Acquired lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1248.505059] env[70013]: DEBUG nova.network.neutron [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing network info cache for port b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1248.771716] env[70013]: DEBUG nova.objects.base [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Object Instance<96a5b996-7b8b-4a3d-a815-5c7dcf1090d6> lazy-loaded attributes: resources,numa_topology {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1248.861034] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 1eea3d47-589d-450b-b5f8-2e763cd1bbd3] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1248.907884] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed14b4d5-8c37-44f3-9481-d2f49cca9985 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.916060] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2518ff-5a68-4c1c-81db-758f148ce47f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.947519] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91be6cf3-d305-4669-9289-0ddb081cc05b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.955961] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72963a54-26a9-4800-b4be-19e44165e0f9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.970050] env[70013]: DEBUG nova.compute.provider_tree [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1249.362408] env[70013]: DEBUG nova.network.neutron [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updated VIF entry in instance network info cache for port b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1249.362715] env[70013]: DEBUG nova.network.neutron [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.365703] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: e11550bf-3284-4ae6-9aa9-34923da589ca] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1249.368098] env[70013]: DEBUG nova.network.neutron [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.472997] env[70013]: DEBUG nova.scheduler.client.report [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1249.866150] env[70013]: DEBUG oslo_concurrency.lockutils [req-ec7c879a-cd0c-4fc5-ba48-794ab8939606 req-faff6da9-b504-4d79-b467-f0c25c77b709 service nova] Releasing lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1249.870030] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 499bb115-146c-48dc-9447-3c01c9549a73] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1249.872477] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1249.977887] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.212s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1250.254016] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.374810] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 8f4d999d-54e9-4987-81f7-27a2b6d4ed17] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1250.377491] env[70013]: DEBUG nova.compute.manager [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=70013) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1250.377639] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.377891] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.485583] env[70013]: DEBUG oslo_concurrency.lockutils [None req-df2e6d06-3001-4f27-a64f-4f191d3ce146 tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.983s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1250.486291] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.233s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.486455] env[70013]: INFO nova.compute.manager [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Unshelving [ 1250.880163] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 38738d80-25ca-4a17-81f1-e0a7bc8c4a44] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1250.882713] env[70013]: DEBUG nova.objects.instance [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'migration_context' on Instance uuid 351f5ef4-ed63-4d5d-b0a1-c029393898ac {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1251.385404] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 8dd3e9c7-3a92-49e6-883c-cda647730f5e] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1251.505640] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1251.526760] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8271ff51-6646-406c-99b1-51f3e2726c67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.534993] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93653c6-ea8e-446b-abd7-3892e94161a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.565509] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bd44a6-4a69-49aa-9a3b-b7cd459b791c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.573724] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad7de06-a527-4c75-849a-8d3247e64546 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.587380] env[70013]: DEBUG nova.compute.provider_tree [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.892573] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: bf138a0d-ea1f-4c93-9c7f-e0e3309597b2] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1252.090504] env[70013]: DEBUG nova.scheduler.client.report [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1252.396017] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: e3b37310-262a-4852-89b4-4d1b1d6f5776] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1252.900947] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: be86d96e-15be-4c39-b05d-59da0668c0d6] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1253.100690] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.723s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.106946] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.601s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1253.107175] env[70013]: DEBUG nova.objects.instance [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'pci_requests' on Instance uuid 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.404690] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 456c4926-2ce4-4ba8-a8a6-ca50650e19f6] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1253.615214] env[70013]: DEBUG nova.objects.instance [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'numa_topology' on Instance uuid 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.908247] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 930ababf-5f2c-4b7e-a24d-7831fbd73d66] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1254.118612] env[70013]: INFO nova.compute.claims [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1254.412143] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: addf8ba9-142c-4458-8586-b06e52118d71] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1254.642324] env[70013]: INFO nova.compute.manager [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Swapping old allocation on dict_keys(['dc0a4c9d-b611-453e-a900-41e280c29b95']) held by migration 40a44e67-2ade-43ac-b4ca-2074cef220fd for instance [ 1254.666757] env[70013]: DEBUG nova.scheduler.client.report [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Overwriting current allocation {'allocations': {'dc0a4c9d-b611-453e-a900-41e280c29b95': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 131}}, 'project_id': '51b2e9e21dcc447d86aa3d2f5f307d4f', 'user_id': '3d6c00fc04b8435388ae5e2652f8e0eb', 'consumer_generation': 1} on consumer 351f5ef4-ed63-4d5d-b0a1-c029393898ac {{(pid=70013) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1254.741798] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.741990] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1254.742225] env[70013]: DEBUG nova.network.neutron [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1254.915352] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: c57e876a-8696-4ce5-b533-7580f484414e] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1255.249022] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b433c4a9-e1c7-416a-8feb-9d0ec7d29619 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.257025] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3260cfc-5260-4cad-81e8-67041773789b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.288940] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d471ab-0496-47d2-b3ea-76be5e82d24a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.296789] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76c55b9-14e8-4343-a371-6b6c11892926 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.310099] env[70013]: DEBUG nova.compute.provider_tree [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1255.418533] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: ec7d0e9a-4a73-4681-9395-a3c963772f62] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1255.465462] env[70013]: DEBUG nova.network.neutron [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [{"id": "461fa18b-f3db-4db6-ba87-fe872af49659", "address": "fa:16:3e:d9:00:2c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461fa18b-f3", "ovs_interfaceid": "461fa18b-f3db-4db6-ba87-fe872af49659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.813721] env[70013]: DEBUG nova.scheduler.client.report [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1255.922887] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 8810bebc-448c-4e9b-9fbb-ce9e24611dc9] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1255.967804] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-351f5ef4-ed63-4d5d-b0a1-c029393898ac" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1255.968394] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1255.969362] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-826f5b47-4f23-4506-b31f-9cf805b6ea12 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.976507] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1255.976507] env[70013]: value = "task-4231614" [ 1255.976507] env[70013]: _type = "Task" [ 1255.976507] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.986828] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231614, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.319035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.212s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1256.347659] env[70013]: INFO nova.network.neutron [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating port b9586fe7-a244-4613-a095-504009e343a0 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1256.426534] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 5b60fca9-7ef3-48e6-8a5c-6c0b3ee711a6] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1256.486570] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231614, 'name': PowerOffVM_Task, 'duration_secs': 0.212259} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.486834] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1256.487515] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1256.487735] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1256.487892] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1256.488084] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1256.488231] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1256.488374] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1256.488574] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1256.488728] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1256.488891] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1256.489066] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1256.489240] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1256.494009] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77a7b82e-0b48-48f2-b139-415ef8345e4b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.509105] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1256.509105] env[70013]: value = "task-4231615" [ 1256.509105] env[70013]: _type = "Task" [ 1256.509105] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.517456] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231615, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.930362] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 9026841c-7bda-41e5-a4ac-03d0d3e37560] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1257.019315] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231615, 'name': ReconfigVM_Task, 'duration_secs': 0.136963} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.020430] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07eca397-7dbd-41fd-86c2-e4090feda04b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.039062] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1257.039305] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1257.039464] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1257.039647] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1257.039792] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1257.039940] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1257.040223] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1257.040384] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1257.040549] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1257.040712] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1257.040884] env[70013]: DEBUG nova.virt.hardware [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1257.041704] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cacf4739-f001-40f1-bd30-4ad5262c8fdf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.047529] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1257.047529] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cf92d6-b73a-aa8d-6bde-bd1a7151bb1d" [ 1257.047529] env[70013]: _type = "Task" [ 1257.047529] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.055858] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cf92d6-b73a-aa8d-6bde-bd1a7151bb1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.434273] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 181238eb-cb0c-4740-9896-ca745b53ebe8] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1257.558603] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cf92d6-b73a-aa8d-6bde-bd1a7151bb1d, 'name': SearchDatastore_Task, 'duration_secs': 0.007699} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.564109] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1257.564459] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5650988b-b954-41a4-9262-090a35e4ecf2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.584760] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1257.584760] env[70013]: value = "task-4231616" [ 1257.584760] env[70013]: _type = "Task" [ 1257.584760] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.593456] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231616, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.935588] env[70013]: DEBUG nova.compute.manager [req-82f13b66-fa64-4d97-86a0-1d6da9c7b840 req-359628f0-1886-4a05-91cd-5199a383c5c1 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-vif-plugged-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1257.935820] env[70013]: DEBUG oslo_concurrency.lockutils [req-82f13b66-fa64-4d97-86a0-1d6da9c7b840 req-359628f0-1886-4a05-91cd-5199a383c5c1 service nova] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1257.936040] env[70013]: DEBUG oslo_concurrency.lockutils [req-82f13b66-fa64-4d97-86a0-1d6da9c7b840 req-359628f0-1886-4a05-91cd-5199a383c5c1 service nova] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1257.936244] env[70013]: DEBUG oslo_concurrency.lockutils [req-82f13b66-fa64-4d97-86a0-1d6da9c7b840 req-359628f0-1886-4a05-91cd-5199a383c5c1 service nova] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1257.936417] env[70013]: DEBUG nova.compute.manager [req-82f13b66-fa64-4d97-86a0-1d6da9c7b840 req-359628f0-1886-4a05-91cd-5199a383c5c1 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] No waiting events found dispatching network-vif-plugged-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1257.936584] env[70013]: WARNING nova.compute.manager [req-82f13b66-fa64-4d97-86a0-1d6da9c7b840 req-359628f0-1886-4a05-91cd-5199a383c5c1 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received unexpected event network-vif-plugged-b9586fe7-a244-4613-a095-504009e343a0 for instance with vm_state shelved_offloaded and task_state spawning. [ 1257.937400] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 1230ca35-1512-4464-85fc-d3b4ab05eac1] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1258.022435] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.022435] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1258.022435] env[70013]: DEBUG nova.network.neutron [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1258.097161] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231616, 'name': ReconfigVM_Task, 'duration_secs': 0.181586} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.097161] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1258.098141] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6415ce-1d0e-4a82-b851-0b607881b927 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.121413] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1258.122068] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c36d196-1012-4300-b080-729c64b470d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.141941] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1258.141941] env[70013]: value = "task-4231617" [ 1258.141941] env[70013]: _type = "Task" [ 1258.141941] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.150977] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231617, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.441033] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 073f3ebf-1672-47fc-b4ed-cdb3651a8ee7] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1258.653237] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231617, 'name': ReconfigVM_Task, 'duration_secs': 0.278958} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.653538] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac/351f5ef4-ed63-4d5d-b0a1-c029393898ac.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1258.654456] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fe659d-b500-496d-ad05-7c36552b62b7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.674377] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7341f36a-5ab7-4046-b9cd-2ea692cd0b8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.696163] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27299d91-5c00-48f7-ba7d-9285c79f2f39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.715031] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ec52fe-5f6f-4cfb-b73d-a7907ea03364 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.723717] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1258.724061] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-effe0bef-831d-4959-97cf-966bb9d97265 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.731247] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1258.731247] env[70013]: value = "task-4231618" [ 1258.731247] env[70013]: _type = "Task" [ 1258.731247] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.744132] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.770930] env[70013]: DEBUG nova.network.neutron [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": "b9586fe7-a244-4613-a095-504009e343a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1258.944158] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 0c00737c-3fd4-4ec6-8b47-1ee64096e8f9] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1259.244707] env[70013]: DEBUG oslo_vmware.api [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231618, 'name': PowerOnVM_Task, 'duration_secs': 0.394233} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.244971] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1259.274094] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1259.301162] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='d639632c625e4f25fd55a6a80999fdd5',container_format='bare',created_at=2025-05-19T11:42:08Z,direct_url=,disk_format='vmdk',id=4da16937-60b6-4c9f-b8ec-afe9cdd73248,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1537044429-shelved',owner='0f5b0bd88f5046f098579c5d59e3f0c3',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2025-05-19T11:42:21Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1259.301424] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1259.301581] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1259.301759] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1259.301902] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1259.302058] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1259.302276] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1259.302428] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1259.302589] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1259.302748] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1259.302919] env[70013]: DEBUG nova.virt.hardware [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1259.303845] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f3518f-61e9-4f25-9e3b-ca344fda2d31 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.312146] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0cbeec-c3ee-4127-adab-1792cc3a8f29 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.327466] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:8a:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9586fe7-a244-4613-a095-504009e343a0', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1259.334751] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1259.334998] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1259.335222] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38cf32c1-f09f-421d-814f-5724d0647d25 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.355623] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1259.355623] env[70013]: value = "task-4231619" [ 1259.355623] env[70013]: _type = "Task" [ 1259.355623] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.363876] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231619, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.447223] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 2070989d-6c90-4eb1-8508-7587ed9659d6] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1259.865706] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231619, 'name': CreateVM_Task, 'duration_secs': 0.296901} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.865918] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1259.866653] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.866821] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1259.867231] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1259.867518] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdc3c857-83d2-4b7b-88bc-d5b4b8f90aeb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.872752] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1259.872752] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]528690d0-7910-6c1a-d401-5f65886bfdd8" [ 1259.872752] env[70013]: _type = "Task" [ 1259.872752] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.880944] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]528690d0-7910-6c1a-d401-5f65886bfdd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.950600] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: b292a985-3d2a-4cf3-a09b-8e72d8d21078] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1259.962710] env[70013]: DEBUG nova.compute.manager [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-changed-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1259.962902] env[70013]: DEBUG nova.compute.manager [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing instance network info cache due to event network-changed-b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1259.963137] env[70013]: DEBUG oslo_concurrency.lockutils [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] Acquiring lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.963280] env[70013]: DEBUG oslo_concurrency.lockutils [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] Acquired lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1259.963432] env[70013]: DEBUG nova.network.neutron [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Refreshing network info cache for port b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1260.256380] env[70013]: INFO nova.compute.manager [None req-0f7f87b1-7e18-40f6-90f0-42213eaf3ea7 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance to original state: 'active' [ 1260.383569] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1260.383861] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Processing image 4da16937-60b6-4c9f-b8ec-afe9cdd73248 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1260.384145] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248/4da16937-60b6-4c9f-b8ec-afe9cdd73248.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1260.384302] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248/4da16937-60b6-4c9f-b8ec-afe9cdd73248.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1260.384485] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1260.384750] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c891eed3-9c3b-4a31-8b4e-2e97e1983e12 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.396021] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1260.396021] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1260.396021] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52b3692f-1780-4e83-b9d8-aea953eecbe2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.400543] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1260.400543] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5272ac64-166a-7b5d-e02b-7662ccfad126" [ 1260.400543] env[70013]: _type = "Task" [ 1260.400543] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.409244] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5272ac64-166a-7b5d-e02b-7662ccfad126, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.454449] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 54d39dfd-6429-4009-8d26-22c1fa46243e] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1260.662568] env[70013]: DEBUG nova.network.neutron [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updated VIF entry in instance network info cache for port b9586fe7-a244-4613-a095-504009e343a0. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1260.662971] env[70013]: DEBUG nova.network.neutron [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [{"id": "b9586fe7-a244-4613-a095-504009e343a0", "address": "fa:16:3e:c7:8a:d5", "network": {"id": "76b3436a-c52c-4eae-8be3-ab248cb31d13", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1841157710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f5b0bd88f5046f098579c5d59e3f0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9586fe7-a2", "ovs_interfaceid": "b9586fe7-a244-4613-a095-504009e343a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1260.911437] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Preparing fetch location {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1260.911748] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Fetch image to [datastore2] OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c/OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c.vmdk {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1260.911920] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Downloading stream optimized image 4da16937-60b6-4c9f-b8ec-afe9cdd73248 to [datastore2] OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c/OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c.vmdk on the data store datastore2 as vApp {{(pid=70013) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1260.912032] env[70013]: DEBUG nova.virt.vmwareapi.images [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Downloading image file data 4da16937-60b6-4c9f-b8ec-afe9cdd73248 to the ESX as VM named 'OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c' {{(pid=70013) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1260.958058] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 0ea84d0d-905c-428d-8abb-2781c817f08f] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1260.980713] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1260.980713] env[70013]: value = "resgroup-9" [ 1260.980713] env[70013]: _type = "ResourcePool" [ 1260.980713] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1260.981058] env[70013]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b14094af-c37f-40b6-8810-cb2498f36080 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.003093] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease: (returnval){ [ 1261.003093] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527cecca-73d4-d85a-505d-fc1002bfb725" [ 1261.003093] env[70013]: _type = "HttpNfcLease" [ 1261.003093] env[70013]: } obtained for vApp import into resource pool (val){ [ 1261.003093] env[70013]: value = "resgroup-9" [ 1261.003093] env[70013]: _type = "ResourcePool" [ 1261.003093] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1261.003383] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the lease: (returnval){ [ 1261.003383] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527cecca-73d4-d85a-505d-fc1002bfb725" [ 1261.003383] env[70013]: _type = "HttpNfcLease" [ 1261.003383] env[70013]: } to be ready. {{(pid=70013) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1261.010121] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1261.010121] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527cecca-73d4-d85a-505d-fc1002bfb725" [ 1261.010121] env[70013]: _type = "HttpNfcLease" [ 1261.010121] env[70013]: } is initializing. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1261.166189] env[70013]: DEBUG oslo_concurrency.lockutils [req-802f694d-57db-4350-9c1a-999c75eb10c3 req-fd783c79-98e4-4a54-a95e-a6af2dae55f3 service nova] Releasing lock "refresh_cache-96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1261.463375] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 8dbef727-13e9-4231-95a8-65d015ee13be] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1261.515617] env[70013]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1261.515617] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527cecca-73d4-d85a-505d-fc1002bfb725" [ 1261.515617] env[70013]: _type = "HttpNfcLease" [ 1261.515617] env[70013]: } is ready. {{(pid=70013) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1261.515908] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1261.515908] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527cecca-73d4-d85a-505d-fc1002bfb725" [ 1261.515908] env[70013]: _type = "HttpNfcLease" [ 1261.515908] env[70013]: }. {{(pid=70013) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1261.516698] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df74a6b-413c-4db1-ae99-1c782cfdaf97 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.524761] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5220e479-c525-a020-f926-adcf90ac828f/disk-0.vmdk from lease info. {{(pid=70013) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1261.524960] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5220e479-c525-a020-f926-adcf90ac828f/disk-0.vmdk. {{(pid=70013) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1261.587618] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-894bd4d4-d862-4326-b5d8-7582f0b9baec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.968104] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 1ed90698-ae2b-4d3a-a3e5-6fb37e2291d6] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1262.012924] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1262.013183] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1262.013391] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1262.013574] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1262.013742] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1262.018466] env[70013]: INFO nova.compute.manager [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Terminating instance [ 1262.470285] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: ab091ef4-9828-438f-8e95-4370bfbd34de] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1262.522492] env[70013]: DEBUG nova.compute.manager [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1262.522813] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1262.523855] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6c1503-2c86-4c8b-aab3-021833ac9b6f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.534496] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1262.536175] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f65dcc75-71d0-4a08-9d3d-e61853fefd67 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.545415] env[70013]: DEBUG oslo_vmware.api [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1262.545415] env[70013]: value = "task-4231621" [ 1262.545415] env[70013]: _type = "Task" [ 1262.545415] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.556479] env[70013]: DEBUG oslo_vmware.api [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231621, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.974096] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 626daf5e-0bfe-4b07-84d8-7c1ddd8ae2df] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1262.989678] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Completed reading data from the image iterator. {{(pid=70013) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1262.989957] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5220e479-c525-a020-f926-adcf90ac828f/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1262.990878] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbde5151-54b2-4518-8158-b8692543856e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.998336] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5220e479-c525-a020-f926-adcf90ac828f/disk-0.vmdk is in state: ready. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1262.998548] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5220e479-c525-a020-f926-adcf90ac828f/disk-0.vmdk. {{(pid=70013) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1262.999468] env[70013]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f11ac85f-ceb2-466e-baaf-b208571713d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.055386] env[70013]: DEBUG oslo_vmware.api [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231621, 'name': PowerOffVM_Task, 'duration_secs': 0.246978} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.055663] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1263.055846] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1263.056068] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd178313-5de3-4999-928b-af5c68bb4a66 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.124994] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1263.125303] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1263.125548] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleting the datastore file [datastore2] 351f5ef4-ed63-4d5d-b0a1-c029393898ac {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1263.125878] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc4f4570-bca6-4935-944e-489c33a8a62f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.132888] env[70013]: DEBUG oslo_vmware.api [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1263.132888] env[70013]: value = "task-4231623" [ 1263.132888] env[70013]: _type = "Task" [ 1263.132888] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.141910] env[70013]: DEBUG oslo_vmware.api [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231623, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.193489] env[70013]: DEBUG oslo_vmware.rw_handles [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5220e479-c525-a020-f926-adcf90ac828f/disk-0.vmdk. {{(pid=70013) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1263.193723] env[70013]: INFO nova.virt.vmwareapi.images [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Downloaded image file data 4da16937-60b6-4c9f-b8ec-afe9cdd73248 [ 1263.194622] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a094fc-0548-4a44-9308-0674006f8b04 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.210585] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1e3da85-672a-4b5c-ac30-a672b0649995 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.235272] env[70013]: INFO nova.virt.vmwareapi.images [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] The imported VM was unregistered [ 1263.237600] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Caching image {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1263.237828] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Creating directory with path [datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1263.238133] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2a04cda-b22d-473c-9ddd-4b9e67c05485 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.336717] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Created directory with path [datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248 {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1263.336926] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c/OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c.vmdk to [datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248/4da16937-60b6-4c9f-b8ec-afe9cdd73248.vmdk. {{(pid=70013) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1263.337213] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-2502a39c-5570-4f12-a853-3668d6febbe2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.345746] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1263.345746] env[70013]: value = "task-4231625" [ 1263.345746] env[70013]: _type = "Task" [ 1263.345746] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.354587] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231625, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.477555] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 472b8218-51c3-492e-96ff-5fa99df4cc4c] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1263.644793] env[70013]: DEBUG oslo_vmware.api [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231623, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316147} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.645184] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1263.645457] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1263.645697] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1263.645908] env[70013]: INFO nova.compute.manager [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1263.646224] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1263.646445] env[70013]: DEBUG nova.compute.manager [-] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1263.646587] env[70013]: DEBUG nova.network.neutron [-] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1263.856849] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231625, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.981689] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 30fcb8c7-0947-426e-9aa5-081446e31c28] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1264.152997] env[70013]: DEBUG nova.compute.manager [req-547eed62-44eb-494a-8730-901f83e38026 req-551c1ea3-f4e8-4045-8efe-aefd2d59f5a5 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Received event network-vif-deleted-461fa18b-f3db-4db6-ba87-fe872af49659 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1264.153252] env[70013]: INFO nova.compute.manager [req-547eed62-44eb-494a-8730-901f83e38026 req-551c1ea3-f4e8-4045-8efe-aefd2d59f5a5 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Neutron deleted interface 461fa18b-f3db-4db6-ba87-fe872af49659; detaching it from the instance and deleting it from the info cache [ 1264.153543] env[70013]: DEBUG nova.network.neutron [req-547eed62-44eb-494a-8730-901f83e38026 req-551c1ea3-f4e8-4045-8efe-aefd2d59f5a5 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.357284] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231625, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.485848] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 965480de-4fe2-48ce-94ad-0838d78239dd] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1264.634430] env[70013]: DEBUG nova.network.neutron [-] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.657082] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f115a92-0b18-49a3-a7ec-a8c651f434f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.667963] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a00603-e4b4-4aee-8476-ae400a56a62d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.694226] env[70013]: DEBUG nova.compute.manager [req-547eed62-44eb-494a-8730-901f83e38026 req-551c1ea3-f4e8-4045-8efe-aefd2d59f5a5 service nova] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Detach interface failed, port_id=461fa18b-f3db-4db6-ba87-fe872af49659, reason: Instance 351f5ef4-ed63-4d5d-b0a1-c029393898ac could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1264.858766] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231625, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.989365] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: eab0c393-4d3d-4659-8225-c82ddadb330b] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1265.138424] env[70013]: INFO nova.compute.manager [-] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Took 1.49 seconds to deallocate network for instance. [ 1265.359025] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231625, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.492456] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 67b90fec-a6fd-471d-a425-82bc63de572a] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1265.645708] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1265.646065] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1265.646334] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1265.671312] env[70013]: INFO nova.scheduler.client.report [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted allocations for instance 351f5ef4-ed63-4d5d-b0a1-c029393898ac [ 1265.859951] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231625, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.376195} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.860284] env[70013]: INFO nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c/OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c.vmdk to [datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248/4da16937-60b6-4c9f-b8ec-afe9cdd73248.vmdk. [ 1265.860488] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Cleaning up location [datastore2] OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1265.860654] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_82903d89-4e0d-49ff-9276-be30cd69649c {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1265.860928] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-340438c9-58d0-46a0-ac4e-7e58ac4023c3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.867376] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1265.867376] env[70013]: value = "task-4231626" [ 1265.867376] env[70013]: _type = "Task" [ 1265.867376] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.877270] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231626, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.996152] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 85488f5e-bd8a-4bcf-8420-744c54176c2e] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1266.180038] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f2061545-9cf6-4ecb-b80d-a19232fac1db tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "351f5ef4-ed63-4d5d-b0a1-c029393898ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.167s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1266.378706] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231626, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033848} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.379806] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1266.379806] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248/4da16937-60b6-4c9f-b8ec-afe9cdd73248.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1266.379806] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248/4da16937-60b6-4c9f-b8ec-afe9cdd73248.vmdk to [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1266.380042] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b40f917-4f62-44f8-9267-5f5c2666dd34 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.388393] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1266.388393] env[70013]: value = "task-4231627" [ 1266.388393] env[70013]: _type = "Task" [ 1266.388393] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.397548] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231627, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.499286] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 2c0e4326-c33a-42bb-b793-a100157b1c03] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1266.899409] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231627, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.002812] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 09544922-6dc4-48b5-8cfd-e91e7f74c13f] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1267.009845] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "3d92b939-18d2-450c-b815-e0266542fbcb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1267.010102] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1267.400463] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231627, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.506441] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: fc473d40-b57b-437e-9511-58a0ba700a69] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1267.512972] env[70013]: DEBUG nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1267.900826] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231627, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.010780] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 63c3b57c-022a-4eee-b215-6dd89142e659] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1268.038659] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1268.038961] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1268.040767] env[70013]: INFO nova.compute.claims [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1268.403329] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231627, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.515116] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 28a56a75-9b30-4121-8252-a9e57287441c] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1268.903059] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231627, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.262725} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.903059] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4da16937-60b6-4c9f-b8ec-afe9cdd73248/4da16937-60b6-4c9f-b8ec-afe9cdd73248.vmdk to [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1268.903059] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ffae61-ba89-447d-ae94-c829c0fd6594 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.925144] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1268.925437] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e70b8ecd-436d-491b-9365-d459a6967d04 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.944979] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1268.944979] env[70013]: value = "task-4231628" [ 1268.944979] env[70013]: _type = "Task" [ 1268.944979] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.953281] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231628, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.185944] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be96b522-77bc-4d93-aa61-8bf7b3e66148 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.194739] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52edb873-2768-4147-9afb-1c6c79d74760 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.235662] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6dbf8d-9f0a-4a55-a385-6ad6846c9f1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.246600] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91132744-3f6c-4161-9b3d-5b9853445426 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.268614] env[70013]: DEBUG nova.compute.provider_tree [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1269.454741] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231628, 'name': ReconfigVM_Task, 'duration_secs': 0.276765} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.455051] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6/96a5b996-7b8b-4a3d-a815-5c7dcf1090d6.vmdk or device None with type streamOptimized {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1269.455703] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe7b6e01-4a3f-4b67-b523-0b80655f3653 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.462302] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1269.462302] env[70013]: value = "task-4231629" [ 1269.462302] env[70013]: _type = "Task" [ 1269.462302] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.471594] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231629, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.772515] env[70013]: DEBUG nova.scheduler.client.report [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1269.973108] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231629, 'name': Rename_Task, 'duration_secs': 0.135936} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.973414] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1269.973687] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac0d8f4f-3ad7-41c5-b65b-2d3efaab61d7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.981106] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1269.981106] env[70013]: value = "task-4231630" [ 1269.981106] env[70013]: _type = "Task" [ 1269.981106] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.989522] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.019466] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1270.277801] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1270.278360] env[70013]: DEBUG nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1270.492022] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231630, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.783829] env[70013]: DEBUG nova.compute.utils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1270.785941] env[70013]: DEBUG nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1270.786164] env[70013]: DEBUG nova.network.neutron [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1270.823862] env[70013]: DEBUG nova.policy [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d6c00fc04b8435388ae5e2652f8e0eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51b2e9e21dcc447d86aa3d2f5f307d4f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1270.994087] env[70013]: DEBUG oslo_vmware.api [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231630, 'name': PowerOnVM_Task, 'duration_secs': 0.52639} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.994376] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1271.096599] env[70013]: DEBUG nova.compute.manager [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1271.097550] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80aa5e70-a247-4158-88ce-826eb90b969c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.114540] env[70013]: DEBUG nova.network.neutron [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Successfully created port: bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1271.287334] env[70013]: DEBUG nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1271.618019] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7407a45e-58eb-4bfe-9d73-a223d75d83aa tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.129s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1272.297919] env[70013]: DEBUG nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1272.325376] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1272.325636] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1272.325796] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1272.325981] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1272.326152] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1272.326300] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1272.326524] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1272.326719] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1272.326896] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1272.327075] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1272.327258] env[70013]: DEBUG nova.virt.hardware [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1272.328351] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f2c91a-ea3c-4b3b-9980-c29a7cb7f2ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.337803] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e952c9-f580-4cc3-a8f1-49131b0b79f2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.484433] env[70013]: DEBUG nova.compute.manager [req-314d0326-6f39-4182-9420-bb2969e4a256 req-94d624f3-30af-46ad-8b71-3e4e72bef7b2 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Received event network-vif-plugged-bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1272.484681] env[70013]: DEBUG oslo_concurrency.lockutils [req-314d0326-6f39-4182-9420-bb2969e4a256 req-94d624f3-30af-46ad-8b71-3e4e72bef7b2 service nova] Acquiring lock "3d92b939-18d2-450c-b815-e0266542fbcb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1272.484880] env[70013]: DEBUG oslo_concurrency.lockutils [req-314d0326-6f39-4182-9420-bb2969e4a256 req-94d624f3-30af-46ad-8b71-3e4e72bef7b2 service nova] Lock "3d92b939-18d2-450c-b815-e0266542fbcb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1272.485090] env[70013]: DEBUG oslo_concurrency.lockutils [req-314d0326-6f39-4182-9420-bb2969e4a256 req-94d624f3-30af-46ad-8b71-3e4e72bef7b2 service nova] Lock "3d92b939-18d2-450c-b815-e0266542fbcb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1272.485205] env[70013]: DEBUG nova.compute.manager [req-314d0326-6f39-4182-9420-bb2969e4a256 req-94d624f3-30af-46ad-8b71-3e4e72bef7b2 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] No waiting events found dispatching network-vif-plugged-bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1272.485370] env[70013]: WARNING nova.compute.manager [req-314d0326-6f39-4182-9420-bb2969e4a256 req-94d624f3-30af-46ad-8b71-3e4e72bef7b2 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Received unexpected event network-vif-plugged-bb5d8627-a5f4-4491-a43e-35304de48645 for instance with vm_state building and task_state spawning. [ 1272.584054] env[70013]: DEBUG nova.network.neutron [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Successfully updated port: bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1273.086477] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.086775] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1273.086840] env[70013]: DEBUG nova.network.neutron [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1273.621567] env[70013]: DEBUG nova.network.neutron [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1273.760512] env[70013]: DEBUG nova.network.neutron [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updating instance_info_cache with network_info: [{"id": "bb5d8627-a5f4-4491-a43e-35304de48645", "address": "fa:16:3e:63:a6:3c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5d8627-a5", "ovs_interfaceid": "bb5d8627-a5f4-4491-a43e-35304de48645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1274.263555] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1274.263870] env[70013]: DEBUG nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Instance network_info: |[{"id": "bb5d8627-a5f4-4491-a43e-35304de48645", "address": "fa:16:3e:63:a6:3c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5d8627-a5", "ovs_interfaceid": "bb5d8627-a5f4-4491-a43e-35304de48645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1274.264445] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:a6:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'abcf0d10-3f3f-45dc-923e-1c78766e2dad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb5d8627-a5f4-4491-a43e-35304de48645', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1274.271987] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1274.272224] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1274.272454] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-92c21e2e-9601-46f8-a7d4-f13741dcffd5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.294122] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1274.294122] env[70013]: value = "task-4231631" [ 1274.294122] env[70013]: _type = "Task" [ 1274.294122] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.306580] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231631, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.513433] env[70013]: DEBUG nova.compute.manager [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Received event network-changed-bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1274.513637] env[70013]: DEBUG nova.compute.manager [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Refreshing instance network info cache due to event network-changed-bb5d8627-a5f4-4491-a43e-35304de48645. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1274.513833] env[70013]: DEBUG oslo_concurrency.lockutils [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] Acquiring lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.514019] env[70013]: DEBUG oslo_concurrency.lockutils [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] Acquired lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1274.514229] env[70013]: DEBUG nova.network.neutron [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Refreshing network info cache for port bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1274.804292] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231631, 'name': CreateVM_Task, 'duration_secs': 0.339995} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.804643] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1274.805150] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.805327] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1274.805643] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1274.805901] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8348da61-f60d-4dda-8855-de22cfda6eda {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.811011] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1274.811011] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c1cad7-fc25-daed-b231-4fbb223e5963" [ 1274.811011] env[70013]: _type = "Task" [ 1274.811011] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.819485] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c1cad7-fc25-daed-b231-4fbb223e5963, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.207689] env[70013]: DEBUG nova.network.neutron [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updated VIF entry in instance network info cache for port bb5d8627-a5f4-4491-a43e-35304de48645. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1275.208063] env[70013]: DEBUG nova.network.neutron [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updating instance_info_cache with network_info: [{"id": "bb5d8627-a5f4-4491-a43e-35304de48645", "address": "fa:16:3e:63:a6:3c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5d8627-a5", "ovs_interfaceid": "bb5d8627-a5f4-4491-a43e-35304de48645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1275.321751] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52c1cad7-fc25-daed-b231-4fbb223e5963, 'name': SearchDatastore_Task, 'duration_secs': 0.02602} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.322090] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1275.322337] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1275.322571] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.322746] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1275.322889] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1275.323175] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f2e3e09-c477-4784-8b95-fac5a1977417 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.332905] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1275.333052] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1275.333725] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-428462ab-f560-4b85-ae29-00a1e633bbf5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.339611] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1275.339611] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dcb976-2a40-b3e5-5ab0-1b2a23ec8eaf" [ 1275.339611] env[70013]: _type = "Task" [ 1275.339611] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.348358] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dcb976-2a40-b3e5-5ab0-1b2a23ec8eaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.710822] env[70013]: DEBUG oslo_concurrency.lockutils [req-2ae20271-4dbf-4908-8477-52e8f1e9b9b7 req-2dff52a0-a647-468b-ae75-71c0eef7a862 service nova] Releasing lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1275.851835] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52dcb976-2a40-b3e5-5ab0-1b2a23ec8eaf, 'name': SearchDatastore_Task, 'duration_secs': 0.010394} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.852656] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf51eb67-7380-4835-97b0-dd6f7719bc39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.857965] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1275.857965] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5299f11e-a6bf-e900-8a74-8a070df10add" [ 1275.857965] env[70013]: _type = "Task" [ 1275.857965] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.866392] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5299f11e-a6bf-e900-8a74-8a070df10add, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.368487] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5299f11e-a6bf-e900-8a74-8a070df10add, 'name': SearchDatastore_Task, 'duration_secs': 0.009456} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.368741] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1276.369011] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 3d92b939-18d2-450c-b815-e0266542fbcb/3d92b939-18d2-450c-b815-e0266542fbcb.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1276.369289] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9790a192-7c79-4843-8c4f-7dd23ec42571 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.376536] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1276.376536] env[70013]: value = "task-4231632" [ 1276.376536] env[70013]: _type = "Task" [ 1276.376536] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.385093] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.886996] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231632, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.389082] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515438} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.389082] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 3d92b939-18d2-450c-b815-e0266542fbcb/3d92b939-18d2-450c-b815-e0266542fbcb.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1277.389467] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1277.389467] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31f954a3-5f20-4cce-a855-710b1c7bed0d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.396347] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1277.396347] env[70013]: value = "task-4231633" [ 1277.396347] env[70013]: _type = "Task" [ 1277.396347] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.404071] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231633, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.907186] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231633, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075339} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.907593] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1277.908243] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f091a1-b0f0-4dee-945c-06c0c953ebb6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.930914] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 3d92b939-18d2-450c-b815-e0266542fbcb/3d92b939-18d2-450c-b815-e0266542fbcb.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1277.931229] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e1b4255-0994-432d-8670-68dfef60ae15 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.953252] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1277.953252] env[70013]: value = "task-4231634" [ 1277.953252] env[70013]: _type = "Task" [ 1277.953252] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.963142] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231634, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.462345] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231634, 'name': ReconfigVM_Task, 'duration_secs': 0.307984} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.462668] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 3d92b939-18d2-450c-b815-e0266542fbcb/3d92b939-18d2-450c-b815-e0266542fbcb.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1278.463350] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e443b15-2075-4728-af05-7dd46b662737 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.472212] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1278.472212] env[70013]: value = "task-4231635" [ 1278.472212] env[70013]: _type = "Task" [ 1278.472212] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.481714] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231635, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.982668] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231635, 'name': Rename_Task, 'duration_secs': 0.151969} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.983072] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1278.983219] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25d68f3f-c0ea-4dba-9866-561531a535e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.990470] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1278.990470] env[70013]: value = "task-4231636" [ 1278.990470] env[70013]: _type = "Task" [ 1278.990470] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.998387] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231636, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.502791] env[70013]: DEBUG oslo_vmware.api [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231636, 'name': PowerOnVM_Task, 'duration_secs': 0.507715} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.503066] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1279.503274] env[70013]: INFO nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Took 7.21 seconds to spawn the instance on the hypervisor. [ 1279.503452] env[70013]: DEBUG nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1279.504331] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96d6330-3085-4460-b080-e6780a63005e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.023913] env[70013]: INFO nova.compute.manager [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Took 12.01 seconds to build instance. [ 1280.526195] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad437b39-5b14-46ef-ba72-b61f95d9deb2 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.516s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1281.163530] env[70013]: DEBUG nova.compute.manager [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Received event network-changed-bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1281.163928] env[70013]: DEBUG nova.compute.manager [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Refreshing instance network info cache due to event network-changed-bb5d8627-a5f4-4491-a43e-35304de48645. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1281.164327] env[70013]: DEBUG oslo_concurrency.lockutils [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] Acquiring lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.164602] env[70013]: DEBUG oslo_concurrency.lockutils [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] Acquired lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1281.164899] env[70013]: DEBUG nova.network.neutron [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Refreshing network info cache for port bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1281.881833] env[70013]: DEBUG nova.network.neutron [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updated VIF entry in instance network info cache for port bb5d8627-a5f4-4491-a43e-35304de48645. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1281.882232] env[70013]: DEBUG nova.network.neutron [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updating instance_info_cache with network_info: [{"id": "bb5d8627-a5f4-4491-a43e-35304de48645", "address": "fa:16:3e:63:a6:3c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5d8627-a5", "ovs_interfaceid": "bb5d8627-a5f4-4491-a43e-35304de48645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1282.385531] env[70013]: DEBUG oslo_concurrency.lockutils [req-b38c6eca-321f-4115-81fd-935d340b278b req-56cf167e-3ed9-4007-b934-ec176ee313bb service nova] Releasing lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1297.201533] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_power_states {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1297.707123] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Getting list of instances from cluster (obj){ [ 1297.707123] env[70013]: value = "domain-c8" [ 1297.707123] env[70013]: _type = "ClusterComputeResource" [ 1297.707123] env[70013]: } {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1297.707912] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c42b67-71ab-48bb-99e6-97d2cd777c3a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.718745] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Got total of 2 instances {{(pid=70013) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1297.718894] env[70013]: WARNING nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] While synchronizing instance power states, found 10 instances in the database and 2 instances on the hypervisor. [ 1297.719098] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid f12b8575-5082-4be9-9bf5-f4279860d19d {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.719429] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 789fffd9-6725-4bf6-9144-dd603b0a521f {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.719698] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid ef64a05a-b514-4c35-81d3-664ae1ad3ff1 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.719943] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.720215] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 2aefaa91-c439-486a-8b19-c6f45f52583f {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.720514] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 39557c50-866a-4510-b840-b1a6a3e3890e {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.720765] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.720975] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 9a61dedd-3764-4bd9-a300-480cc7d14a21 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.721194] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.721401] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Triggering sync for uuid 3d92b939-18d2-450c-b815-e0266542fbcb {{(pid=70013) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1297.721770] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "f12b8575-5082-4be9-9bf5-f4279860d19d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.721981] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.722304] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "789fffd9-6725-4bf6-9144-dd603b0a521f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.722500] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.722771] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.722959] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.723220] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.723402] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.723642] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "2aefaa91-c439-486a-8b19-c6f45f52583f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.723819] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.724067] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "39557c50-866a-4510-b840-b1a6a3e3890e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.724277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "39557c50-866a-4510-b840-b1a6a3e3890e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.724650] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.724837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.725092] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.725303] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.725599] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.725791] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.726110] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "3d92b939-18d2-450c-b815-e0266542fbcb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.726312] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.726642] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-519450a2-f917-42a3-b1f1-8a1e63994018 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.728589] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2af54560-f16e-45fc-b247-1f307e7b8894 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.730201] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e77dc52-b67b-4e7d-adbd-bb3e7c64f3d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.731785] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b752bd49-a2aa-44b2-bfba-50686728d599 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.733276] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af9c9dfc-160b-4876-b9dc-11f5b1beb952 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.735870] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9a2dc58-765f-409a-8e18-444dfe27a172 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.738049] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76f681ac-ef55-4499-a268-f688274a8d78 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.740196] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-561d4c16-0f53-49cb-9b86-225cf1317952 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.743099] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aea5989-bf9a-4879-b5c2-7324410218ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.747413] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23eb02f1-e44b-4f89-8ff1-9b3c3d85e0d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.768433] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d479739-359b-43a3-8f29-33754a6f2ed7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.785701] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde8be45-0191-4b90-ad37-3f70d84de750 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.798414] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2ee049-c4d0-431c-a60a-9459a8a9f9da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.811437] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157c2334-183f-48f5-8ddd-dfc316281aa6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.827845] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d602279d-21db-4662-90c4-1e28a2167bca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.840554] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20717f0f-a012-49da-9521-95f4871ab9fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.865699] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770f79d9-0717-47bb-bcfe-3f1668d37993 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.881804] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73327e11-b227-4fde-a8ff-9d88814e03e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.956976] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "f12b8575-5082-4be9-9bf5-f4279860d19d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.235s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.353683] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.628s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.394550] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.668s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.396178] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "789fffd9-6725-4bf6-9144-dd603b0a521f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.674s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.415859] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "ef64a05a-b514-4c35-81d3-664ae1ad3ff1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.693s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.418351] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "39557c50-866a-4510-b840-b1a6a3e3890e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.694s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.421816] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "2aefaa91-c439-486a-8b19-c6f45f52583f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.698s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.440467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "9137a6bc-8f1e-47b7-b1ec-5b4035d36fff" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.717s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.458058] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "9a61dedd-3764-4bd9-a300-480cc7d14a21" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.733s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.458283] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.733s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1301.865802] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.339803] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.339976] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1302.340152] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.842977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1302.843271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1302.843426] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1302.843588] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1302.844600] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97115952-555a-49b6-9d4f-50b6b732cbd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.853507] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb71bd7-d059-4015-9f2a-bb522ea9c101 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.867833] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772f7838-2f02-4f7e-ab11-d0a881823e79 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.874739] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311c9a06-ddcf-4a4e-8e74-08f94a53a3bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.903819] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180268MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1302.904026] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1302.904199] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1303.936356] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.936651] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.936651] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.936749] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.936855] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.936969] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.937100] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.937216] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.937330] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.937444] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 3d92b939-18d2-450c-b815-e0266542fbcb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1303.937642] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1303.937781] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1304.055728] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe731c9-1a0a-4f0f-a951-692c98949f74 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.064040] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25e2ef4-e999-4c25-b31d-173f410c7eb4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.095044] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19845d80-4ed9-48df-bfd8-b088e79d24bd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.102797] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d57b95-dbc9-427d-8375-614361289cc7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.116411] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1304.619592] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1305.124427] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1305.124810] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.220s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1306.119693] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.120139] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.120139] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.120318] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.995472] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1306.995913] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1306.995982] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1306.996188] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1306.996362] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1306.999835] env[70013]: INFO nova.compute.manager [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Terminating instance [ 1307.503819] env[70013]: DEBUG nova.compute.manager [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1307.504081] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1307.505452] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2735623-9f7f-40ee-a45d-0f8061b24260 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.513225] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1307.513472] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac71b19f-1ed9-465d-855b-c4817d779476 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.519511] env[70013]: DEBUG oslo_vmware.api [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1307.519511] env[70013]: value = "task-4231637" [ 1307.519511] env[70013]: _type = "Task" [ 1307.519511] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.527822] env[70013]: DEBUG oslo_vmware.api [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231637, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.031333] env[70013]: DEBUG oslo_vmware.api [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231637, 'name': PowerOffVM_Task, 'duration_secs': 0.170033} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.031702] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1308.031799] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1308.032026] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df39ceac-e369-4857-a127-f7739738d44c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.099224] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1308.099474] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1308.099639] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleting the datastore file [datastore2] 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1308.099931] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-848ed6aa-d197-4219-ac39-5d746ae33bb8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.107376] env[70013]: DEBUG oslo_vmware.api [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for the task: (returnval){ [ 1308.107376] env[70013]: value = "task-4231639" [ 1308.107376] env[70013]: _type = "Task" [ 1308.107376] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.115740] env[70013]: DEBUG oslo_vmware.api [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231639, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.339817] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1308.618372] env[70013]: DEBUG oslo_vmware.api [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Task: {'id': task-4231639, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123928} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.618811] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1308.619126] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1308.619421] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1308.619712] env[70013]: INFO nova.compute.manager [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1308.620144] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1308.620404] env[70013]: DEBUG nova.compute.manager [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1308.620628] env[70013]: DEBUG nova.network.neutron [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1309.138709] env[70013]: DEBUG nova.compute.manager [req-1321f5b6-7e02-4f01-90c6-7200103ebdc2 req-9a794e2a-aebb-435d-8465-fba8aaad3561 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Received event network-vif-deleted-b9586fe7-a244-4613-a095-504009e343a0 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1309.138964] env[70013]: INFO nova.compute.manager [req-1321f5b6-7e02-4f01-90c6-7200103ebdc2 req-9a794e2a-aebb-435d-8465-fba8aaad3561 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Neutron deleted interface b9586fe7-a244-4613-a095-504009e343a0; detaching it from the instance and deleting it from the info cache [ 1309.139156] env[70013]: DEBUG nova.network.neutron [req-1321f5b6-7e02-4f01-90c6-7200103ebdc2 req-9a794e2a-aebb-435d-8465-fba8aaad3561 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1309.590017] env[70013]: DEBUG nova.network.neutron [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1309.641857] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2a7fee1-6db2-4363-8a25-45e9995dca01 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.656115] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da7604e-a952-4537-a2da-d71c59d4aae8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.693829] env[70013]: DEBUG nova.compute.manager [req-1321f5b6-7e02-4f01-90c6-7200103ebdc2 req-9a794e2a-aebb-435d-8465-fba8aaad3561 service nova] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Detach interface failed, port_id=b9586fe7-a244-4613-a095-504009e343a0, reason: Instance 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1310.093625] env[70013]: INFO nova.compute.manager [-] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Took 1.47 seconds to deallocate network for instance. [ 1310.600352] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1310.600735] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1310.600883] env[70013]: DEBUG nova.objects.instance [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lazy-loading 'resources' on Instance uuid 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1311.235243] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b310032-a4a1-4367-bb59-ff0230a06485 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.243104] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259ec4f4-fdd8-48a0-b69a-69e5277ec292 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.274110] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008d545d-abf4-4cb4-a995-3a4e44920c8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.282179] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f230709f-aaea-43b7-bbe8-ea795ab01e04 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.295939] env[70013]: DEBUG nova.compute.provider_tree [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1311.799792] env[70013]: DEBUG nova.scheduler.client.report [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1312.304930] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1312.324871] env[70013]: INFO nova.scheduler.client.report [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Deleted allocations for instance 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6 [ 1312.834792] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ad77090d-55f4-4c67-9d6a-9fdcd9785f4c tempest-AttachVolumeShelveTestJSON-1969238173 tempest-AttachVolumeShelveTestJSON-1969238173-project-member] Lock "96a5b996-7b8b-4a3d-a815-5c7dcf1090d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.839s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1317.611852] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "3d92b939-18d2-450c-b815-e0266542fbcb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1317.612185] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1317.612336] env[70013]: DEBUG nova.compute.manager [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1317.613270] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae2f75f-a407-42c1-b5e7-e24d2ab9804f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.620614] env[70013]: DEBUG nova.compute.manager [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1317.621188] env[70013]: DEBUG nova.objects.instance [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'flavor' on Instance uuid 3d92b939-18d2-450c-b815-e0266542fbcb {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1318.629538] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1318.630239] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03b2f655-cb0f-4edb-9db9-8e95283ade27 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.639032] env[70013]: DEBUG oslo_vmware.api [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1318.639032] env[70013]: value = "task-4231641" [ 1318.639032] env[70013]: _type = "Task" [ 1318.639032] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.647471] env[70013]: DEBUG oslo_vmware.api [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231641, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.148993] env[70013]: DEBUG oslo_vmware.api [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231641, 'name': PowerOffVM_Task, 'duration_secs': 0.189898} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.149339] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1319.149541] env[70013]: DEBUG nova.compute.manager [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1319.150617] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15730193-ef0d-43f6-89bc-74c40e0199f4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.665514] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b319f407-7735-4fec-bb78-a2817bbfb180 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.053s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1320.013643] env[70013]: DEBUG nova.objects.instance [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'flavor' on Instance uuid 3d92b939-18d2-450c-b815-e0266542fbcb {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1320.518584] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1320.518768] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1320.518947] env[70013]: DEBUG nova.network.neutron [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1320.520036] env[70013]: DEBUG nova.objects.instance [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'info_cache' on Instance uuid 3d92b939-18d2-450c-b815-e0266542fbcb {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1321.022460] env[70013]: DEBUG nova.objects.base [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Object Instance<3d92b939-18d2-450c-b815-e0266542fbcb> lazy-loaded attributes: flavor,info_cache {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1321.759353] env[70013]: DEBUG nova.network.neutron [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updating instance_info_cache with network_info: [{"id": "bb5d8627-a5f4-4491-a43e-35304de48645", "address": "fa:16:3e:63:a6:3c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5d8627-a5", "ovs_interfaceid": "bb5d8627-a5f4-4491-a43e-35304de48645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1322.262413] env[70013]: DEBUG oslo_concurrency.lockutils [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1322.596863] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1322.597296] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1323.099807] env[70013]: DEBUG nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1323.268294] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1323.268629] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e62aa5bc-54a5-488c-8db5-cd266f8647ac {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.276885] env[70013]: DEBUG oslo_vmware.api [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1323.276885] env[70013]: value = "task-4231642" [ 1323.276885] env[70013]: _type = "Task" [ 1323.276885] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.285698] env[70013]: DEBUG oslo_vmware.api [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231642, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.626645] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1323.626941] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1323.628602] env[70013]: INFO nova.compute.claims [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1323.786833] env[70013]: DEBUG oslo_vmware.api [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231642, 'name': PowerOnVM_Task, 'duration_secs': 0.354794} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.787266] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1323.787356] env[70013]: DEBUG nova.compute.manager [None req-89b3d40e-4844-4a18-9d5e-d2c6fcf89fe9 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1323.788106] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ea79d3-a249-494e-9f8b-488176012cde {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.769893] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3765d44-9876-416e-85c6-41034eb6a156 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.778495] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3191cc4b-10db-4c76-9e9f-1739bf2761d5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.809473] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433309cb-3d87-44f8-9a12-910b0b451fd6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.817279] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c7c3cd-fe72-493e-879d-d5749ad27364 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.830580] env[70013]: DEBUG nova.compute.provider_tree [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1324.839437] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd3173f-1f1f-402f-aabd-0ec6acf032a5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.847500] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a38e39aa-5fed-49ac-8e60-4aeaa942a891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Suspending the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1324.847500] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-51520efa-b09e-4cec-8606-6f04d1669c22 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.856040] env[70013]: DEBUG oslo_vmware.api [None req-a38e39aa-5fed-49ac-8e60-4aeaa942a891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1324.856040] env[70013]: value = "task-4231643" [ 1324.856040] env[70013]: _type = "Task" [ 1324.856040] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.863416] env[70013]: DEBUG oslo_vmware.api [None req-a38e39aa-5fed-49ac-8e60-4aeaa942a891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231643, 'name': SuspendVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.335770] env[70013]: DEBUG nova.scheduler.client.report [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1325.366584] env[70013]: DEBUG oslo_vmware.api [None req-a38e39aa-5fed-49ac-8e60-4aeaa942a891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231643, 'name': SuspendVM_Task} progress is 66%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.841936] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.215s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1325.842580] env[70013]: DEBUG nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1325.865943] env[70013]: DEBUG oslo_vmware.api [None req-a38e39aa-5fed-49ac-8e60-4aeaa942a891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231643, 'name': SuspendVM_Task, 'duration_secs': 0.613869} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.866225] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-a38e39aa-5fed-49ac-8e60-4aeaa942a891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Suspended the VM {{(pid=70013) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1325.866403] env[70013]: DEBUG nova.compute.manager [None req-a38e39aa-5fed-49ac-8e60-4aeaa942a891 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1325.867261] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c2d493-4ec6-42c6-b254-c3456dad65d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.348367] env[70013]: DEBUG nova.compute.utils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1326.350883] env[70013]: DEBUG nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1326.351242] env[70013]: DEBUG nova.network.neutron [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1326.398515] env[70013]: DEBUG nova.policy [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a78bb398602948da9be11bd7ba85146d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0670a7ef480b4d39b532968ac97bbe33', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1326.672813] env[70013]: DEBUG nova.network.neutron [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Successfully created port: 0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1326.855182] env[70013]: DEBUG nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1327.179110] env[70013]: INFO nova.compute.manager [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Resuming [ 1327.179765] env[70013]: DEBUG nova.objects.instance [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'flavor' on Instance uuid 3d92b939-18d2-450c-b815-e0266542fbcb {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1327.865806] env[70013]: DEBUG nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1327.891072] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1327.891334] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1327.891494] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1327.891684] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1327.891828] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1327.891973] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1327.892213] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1327.892372] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1327.892630] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1327.892816] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1327.892986] env[70013]: DEBUG nova.virt.hardware [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1327.893897] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4592418-a69c-4966-86a2-86c209834b8d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.902250] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c16c49-6c6f-4a5d-b087-6ee7e11a8ca5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.043044] env[70013]: DEBUG nova.compute.manager [req-944f3021-95f5-4e00-87e7-be9b71578df4 req-9397416a-f483-4e77-b55d-0ff497786b4a service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Received event network-vif-plugged-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1328.043272] env[70013]: DEBUG oslo_concurrency.lockutils [req-944f3021-95f5-4e00-87e7-be9b71578df4 req-9397416a-f483-4e77-b55d-0ff497786b4a service nova] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1328.043480] env[70013]: DEBUG oslo_concurrency.lockutils [req-944f3021-95f5-4e00-87e7-be9b71578df4 req-9397416a-f483-4e77-b55d-0ff497786b4a service nova] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1328.043651] env[70013]: DEBUG oslo_concurrency.lockutils [req-944f3021-95f5-4e00-87e7-be9b71578df4 req-9397416a-f483-4e77-b55d-0ff497786b4a service nova] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1328.043818] env[70013]: DEBUG nova.compute.manager [req-944f3021-95f5-4e00-87e7-be9b71578df4 req-9397416a-f483-4e77-b55d-0ff497786b4a service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] No waiting events found dispatching network-vif-plugged-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1328.043980] env[70013]: WARNING nova.compute.manager [req-944f3021-95f5-4e00-87e7-be9b71578df4 req-9397416a-f483-4e77-b55d-0ff497786b4a service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Received unexpected event network-vif-plugged-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 for instance with vm_state building and task_state spawning. [ 1328.126427] env[70013]: DEBUG nova.network.neutron [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Successfully updated port: 0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1328.630165] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.630165] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1328.630165] env[70013]: DEBUG nova.network.neutron [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1328.690337] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.690522] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquired lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1328.690707] env[70013]: DEBUG nova.network.neutron [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1329.162361] env[70013]: DEBUG nova.network.neutron [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1329.354322] env[70013]: DEBUG nova.network.neutron [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating instance_info_cache with network_info: [{"id": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "address": "fa:16:3e:f6:03:32", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0641d8ef-98", "ovs_interfaceid": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.452957] env[70013]: DEBUG nova.network.neutron [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updating instance_info_cache with network_info: [{"id": "bb5d8627-a5f4-4491-a43e-35304de48645", "address": "fa:16:3e:63:a6:3c", "network": {"id": "2529a457-488b-466a-9cad-bb7e2b981870", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1157949858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51b2e9e21dcc447d86aa3d2f5f307d4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5d8627-a5", "ovs_interfaceid": "bb5d8627-a5f4-4491-a43e-35304de48645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.857844] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1329.858179] env[70013]: DEBUG nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Instance network_info: |[{"id": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "address": "fa:16:3e:f6:03:32", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0641d8ef-98", "ovs_interfaceid": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1329.858654] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:03:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0641d8ef-98b5-48f3-a1a8-6bdafa0154d4', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1329.866344] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Creating folder: Project (0670a7ef480b4d39b532968ac97bbe33). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1329.866618] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6375e862-80c3-427f-beb3-485f1d459b7c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.879441] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Created folder: Project (0670a7ef480b4d39b532968ac97bbe33) in parent group-v836999. [ 1329.879633] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Creating folder: Instances. Parent ref: group-v837241. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1329.879869] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b3fde8a-69fe-4000-987c-d5f67610e76e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.889257] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Created folder: Instances in parent group-v837241. [ 1329.889491] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1329.889685] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1329.889885] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51dfbecf-c708-416c-9a0b-f3be967c8c08 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.910228] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1329.910228] env[70013]: value = "task-4231646" [ 1329.910228] env[70013]: _type = "Task" [ 1329.910228] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.918217] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231646, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.956566] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Releasing lock "refresh_cache-3d92b939-18d2-450c-b815-e0266542fbcb" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1329.957598] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6990bac3-065e-4156-a300-df34d20f6699 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.964744] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Resuming the VM {{(pid=70013) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1329.965052] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-223eb0e5-2cf2-4a4b-bf59-02e635af47fd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.971829] env[70013]: DEBUG oslo_vmware.api [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1329.971829] env[70013]: value = "task-4231647" [ 1329.971829] env[70013]: _type = "Task" [ 1329.971829] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.980533] env[70013]: DEBUG oslo_vmware.api [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.072280] env[70013]: DEBUG nova.compute.manager [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Received event network-changed-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1330.072497] env[70013]: DEBUG nova.compute.manager [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Refreshing instance network info cache due to event network-changed-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1330.073193] env[70013]: DEBUG oslo_concurrency.lockutils [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] Acquiring lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.073193] env[70013]: DEBUG oslo_concurrency.lockutils [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] Acquired lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1330.073193] env[70013]: DEBUG nova.network.neutron [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Refreshing network info cache for port 0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1330.421069] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231646, 'name': CreateVM_Task, 'duration_secs': 0.319672} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.421435] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1330.429565] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.429797] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1330.430084] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1330.430371] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-732a400c-59f2-4781-9f11-c52bf41a554b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.436727] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1330.436727] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ec33d8-812c-0ba0-277d-5e5c044b1ca1" [ 1330.436727] env[70013]: _type = "Task" [ 1330.436727] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.445988] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ec33d8-812c-0ba0-277d-5e5c044b1ca1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.482067] env[70013]: DEBUG oslo_vmware.api [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231647, 'name': PowerOnVM_Task} progress is 100%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.792389] env[70013]: DEBUG nova.network.neutron [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updated VIF entry in instance network info cache for port 0641d8ef-98b5-48f3-a1a8-6bdafa0154d4. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1330.792768] env[70013]: DEBUG nova.network.neutron [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating instance_info_cache with network_info: [{"id": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "address": "fa:16:3e:f6:03:32", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0641d8ef-98", "ovs_interfaceid": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1330.947348] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ec33d8-812c-0ba0-277d-5e5c044b1ca1, 'name': SearchDatastore_Task, 'duration_secs': 0.047265} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.947647] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1330.947878] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1330.948140] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.948293] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1330.948474] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1330.948735] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-079f32b8-71f3-42a9-9be4-a7541883e3cd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.958251] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1330.958426] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1330.959224] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c55f788-4bd2-4042-9106-b56f643d667b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.965541] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1330.965541] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526358fe-108b-5728-2888-73bcfedc77c1" [ 1330.965541] env[70013]: _type = "Task" [ 1330.965541] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.974022] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526358fe-108b-5728-2888-73bcfedc77c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.983120] env[70013]: DEBUG oslo_vmware.api [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231647, 'name': PowerOnVM_Task, 'duration_secs': 0.535053} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.983368] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Resumed the VM {{(pid=70013) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1330.983547] env[70013]: DEBUG nova.compute.manager [None req-f5f69222-9409-4f04-99fc-fbdaa97d1355 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1330.984326] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a6c6aa-753e-49a8-b5d8-b1bda9cfd755 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.295513] env[70013]: DEBUG oslo_concurrency.lockutils [req-2884dd90-efd0-4ef9-9f31-3e6100eba2a4 req-dd81ce13-9671-4670-992c-d2e0fec6c42d service nova] Releasing lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1331.477170] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526358fe-108b-5728-2888-73bcfedc77c1, 'name': SearchDatastore_Task, 'duration_secs': 0.010918} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.477972] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87a76eae-0324-44bd-8589-147febafd27c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.483795] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1331.483795] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523cbfdf-a7ad-8fb0-a5b6-055c4f68979d" [ 1331.483795] env[70013]: _type = "Task" [ 1331.483795] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.491752] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523cbfdf-a7ad-8fb0-a5b6-055c4f68979d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.837701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "3d92b939-18d2-450c-b815-e0266542fbcb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1331.838087] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1331.838261] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "3d92b939-18d2-450c-b815-e0266542fbcb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1331.838408] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1331.838596] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1331.841073] env[70013]: INFO nova.compute.manager [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Terminating instance [ 1331.995065] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523cbfdf-a7ad-8fb0-a5b6-055c4f68979d, 'name': SearchDatastore_Task, 'duration_secs': 0.010619} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.995308] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1331.995565] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8b74079f-e9d3-4212-8b7c-31f59e75ca13/8b74079f-e9d3-4212-8b7c-31f59e75ca13.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1331.995833] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2618c95-1ada-4ff6-aa9d-2c6c2f4e7175 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.003748] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1332.003748] env[70013]: value = "task-4231648" [ 1332.003748] env[70013]: _type = "Task" [ 1332.003748] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.011764] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231648, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.345087] env[70013]: DEBUG nova.compute.manager [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1332.345497] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1332.346642] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7725804-95ba-4d45-abf5-18df50fecdd0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.356402] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1332.356962] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-263919f0-0810-479a-9c84-7cbd15f079ae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.365644] env[70013]: DEBUG oslo_vmware.api [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1332.365644] env[70013]: value = "task-4231649" [ 1332.365644] env[70013]: _type = "Task" [ 1332.365644] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.377976] env[70013]: DEBUG oslo_vmware.api [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.515083] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231648, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.880739] env[70013]: DEBUG oslo_vmware.api [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231649, 'name': PowerOffVM_Task, 'duration_secs': 0.399429} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.881252] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1332.881612] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1332.882055] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25de2a7f-f105-454d-88e8-a60405ae5fa4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.977044] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1332.977274] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1332.977402] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleting the datastore file [datastore2] 3d92b939-18d2-450c-b815-e0266542fbcb {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1332.977672] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be2a9cba-5bcd-4d35-b117-77bbb42b5335 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.984658] env[70013]: DEBUG oslo_vmware.api [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for the task: (returnval){ [ 1332.984658] env[70013]: value = "task-4231651" [ 1332.984658] env[70013]: _type = "Task" [ 1332.984658] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.993244] env[70013]: DEBUG oslo_vmware.api [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.013923] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231648, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527377} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.014214] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8b74079f-e9d3-4212-8b7c-31f59e75ca13/8b74079f-e9d3-4212-8b7c-31f59e75ca13.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1333.014472] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1333.014813] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d99bb3b-1543-4e2b-8b7d-ed5e8140d72f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.024337] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1333.024337] env[70013]: value = "task-4231652" [ 1333.024337] env[70013]: _type = "Task" [ 1333.024337] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.034534] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231652, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.495561] env[70013]: DEBUG oslo_vmware.api [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Task: {'id': task-4231651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170647} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.495893] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1333.496213] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1333.496520] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1333.496809] env[70013]: INFO nova.compute.manager [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1333.497103] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1333.497303] env[70013]: DEBUG nova.compute.manager [-] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1333.497409] env[70013]: DEBUG nova.network.neutron [-] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1333.534843] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231652, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.400881} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.535244] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1333.535933] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762d6d81-d663-4af5-8225-e07c8fb7570b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.559646] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 8b74079f-e9d3-4212-8b7c-31f59e75ca13/8b74079f-e9d3-4212-8b7c-31f59e75ca13.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1333.562302] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e39ca9a0-8832-4993-b9de-ca209e9bcaee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.583423] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1333.583423] env[70013]: value = "task-4231653" [ 1333.583423] env[70013]: _type = "Task" [ 1333.583423] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.592783] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231653, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.005831] env[70013]: DEBUG nova.compute.manager [req-4cc919a5-9fc6-4cfc-aa37-80bd51f29a41 req-ab10b08a-b818-44f9-9596-a6614000e885 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Received event network-vif-deleted-bb5d8627-a5f4-4491-a43e-35304de48645 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1334.006233] env[70013]: INFO nova.compute.manager [req-4cc919a5-9fc6-4cfc-aa37-80bd51f29a41 req-ab10b08a-b818-44f9-9596-a6614000e885 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Neutron deleted interface bb5d8627-a5f4-4491-a43e-35304de48645; detaching it from the instance and deleting it from the info cache [ 1334.006854] env[70013]: DEBUG nova.network.neutron [req-4cc919a5-9fc6-4cfc-aa37-80bd51f29a41 req-ab10b08a-b818-44f9-9596-a6614000e885 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1334.093118] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231653, 'name': ReconfigVM_Task, 'duration_secs': 0.467244} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.093416] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 8b74079f-e9d3-4212-8b7c-31f59e75ca13/8b74079f-e9d3-4212-8b7c-31f59e75ca13.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1334.094070] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-488d770d-e324-4bb8-a4c1-5559ac2ba1cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.102246] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1334.102246] env[70013]: value = "task-4231654" [ 1334.102246] env[70013]: _type = "Task" [ 1334.102246] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.111026] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231654, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.477114] env[70013]: DEBUG nova.network.neutron [-] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1334.510710] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b0872ef-ddec-4c95-9bae-16b259e0c018 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.520898] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb64213-6f25-4968-9c9c-46c8a9241101 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.546200] env[70013]: DEBUG nova.compute.manager [req-4cc919a5-9fc6-4cfc-aa37-80bd51f29a41 req-ab10b08a-b818-44f9-9596-a6614000e885 service nova] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Detach interface failed, port_id=bb5d8627-a5f4-4491-a43e-35304de48645, reason: Instance 3d92b939-18d2-450c-b815-e0266542fbcb could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1334.612336] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231654, 'name': Rename_Task, 'duration_secs': 0.14762} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.612616] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1334.612882] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d803ed46-c1d3-4cd4-97ec-e1668e9cc1ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.620564] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1334.620564] env[70013]: value = "task-4231655" [ 1334.620564] env[70013]: _type = "Task" [ 1334.620564] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.630272] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.979859] env[70013]: INFO nova.compute.manager [-] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Took 1.48 seconds to deallocate network for instance. [ 1335.130983] env[70013]: DEBUG oslo_vmware.api [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231655, 'name': PowerOnVM_Task, 'duration_secs': 0.510047} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.131312] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1335.131519] env[70013]: INFO nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Took 7.27 seconds to spawn the instance on the hypervisor. [ 1335.131699] env[70013]: DEBUG nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1335.132526] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc288f7-0983-48cd-b6f4-05f2b0e68f4a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.486578] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1335.486813] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1335.487046] env[70013]: DEBUG nova.objects.instance [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lazy-loading 'resources' on Instance uuid 3d92b939-18d2-450c-b815-e0266542fbcb {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1335.651027] env[70013]: INFO nova.compute.manager [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Took 12.04 seconds to build instance. [ 1336.122261] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c9f5f0-12ce-4a73-b807-df4d84b2241b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.130076] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6151af-9112-4a6c-91ee-59d513bd4fce {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.161056] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b251bd18-d200-4244-8a49-3a6097414365 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.563s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1336.161408] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5d1339-567c-41ca-bece-60fb6a01b819 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.169649] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5c8d46-9148-4f3e-8b4e-fcfb24bd76e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.185788] env[70013]: DEBUG nova.compute.provider_tree [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1336.689398] env[70013]: DEBUG nova.scheduler.client.report [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1336.718174] env[70013]: DEBUG nova.compute.manager [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Received event network-changed-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1336.718174] env[70013]: DEBUG nova.compute.manager [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Refreshing instance network info cache due to event network-changed-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1336.718174] env[70013]: DEBUG oslo_concurrency.lockutils [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] Acquiring lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1336.718174] env[70013]: DEBUG oslo_concurrency.lockutils [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] Acquired lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1336.718174] env[70013]: DEBUG nova.network.neutron [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Refreshing network info cache for port 0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1337.194949] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1337.214288] env[70013]: INFO nova.scheduler.client.report [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Deleted allocations for instance 3d92b939-18d2-450c-b815-e0266542fbcb [ 1337.442328] env[70013]: DEBUG nova.network.neutron [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updated VIF entry in instance network info cache for port 0641d8ef-98b5-48f3-a1a8-6bdafa0154d4. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1337.442704] env[70013]: DEBUG nova.network.neutron [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating instance_info_cache with network_info: [{"id": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "address": "fa:16:3e:f6:03:32", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0641d8ef-98", "ovs_interfaceid": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.722146] env[70013]: DEBUG oslo_concurrency.lockutils [None req-49759526-50d3-465a-97fc-3a94ef452e10 tempest-ServerActionsTestJSON-71305913 tempest-ServerActionsTestJSON-71305913-project-member] Lock "3d92b939-18d2-450c-b815-e0266542fbcb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.884s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1337.946283] env[70013]: DEBUG oslo_concurrency.lockutils [req-4ec02679-93c5-4027-b70c-0983f380693f req-8c97286d-67e2-4ce3-9e5c-c4cd5623adeb service nova] Releasing lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1343.284826] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "8d90b797-118d-45cb-9422-7137b7fadbe2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1343.284826] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1343.787604] env[70013]: DEBUG nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1344.310476] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1344.310804] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1344.312388] env[70013]: INFO nova.compute.claims [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1345.453256] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d74e5e4-bdbb-46c7-85e2-cbbf87d6d507 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.461376] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09b1572-98f5-4c8c-b078-28c9c7fa7cf8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.491738] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509f226e-c0b6-4e8b-a664-98813b5f3f33 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.499732] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03ad9d7-e4ec-40a9-be0a-297f9db484ec {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.513081] env[70013]: DEBUG nova.compute.provider_tree [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1346.016636] env[70013]: DEBUG nova.scheduler.client.report [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1346.521538] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.211s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1346.522205] env[70013]: DEBUG nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1347.026950] env[70013]: DEBUG nova.compute.utils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1347.028436] env[70013]: DEBUG nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1347.028639] env[70013]: DEBUG nova.network.neutron [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1347.076136] env[70013]: DEBUG nova.policy [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ade87067e834404adc174783ff7b03d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66804721098f48d69e216807291268d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1347.375961] env[70013]: DEBUG nova.network.neutron [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Successfully created port: 09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1347.531829] env[70013]: DEBUG nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1348.541969] env[70013]: DEBUG nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1348.567539] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1348.567792] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1348.567956] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1348.568160] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1348.568309] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1348.568454] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1348.568703] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1348.568877] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1348.569060] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1348.569234] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1348.569408] env[70013]: DEBUG nova.virt.hardware [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1348.570348] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027e7fb9-97ab-4506-8454-e26d88dc6762 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.579158] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77ba0d7-ad0c-4ce1-94df-efa5fafd24f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.745044] env[70013]: DEBUG nova.compute.manager [req-77ff7b41-f18f-42ac-8625-c2375117b229 req-2598e5af-360b-44c2-bf0a-1c93a6d74910 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Received event network-vif-plugged-09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1348.745044] env[70013]: DEBUG oslo_concurrency.lockutils [req-77ff7b41-f18f-42ac-8625-c2375117b229 req-2598e5af-360b-44c2-bf0a-1c93a6d74910 service nova] Acquiring lock "8d90b797-118d-45cb-9422-7137b7fadbe2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1348.745044] env[70013]: DEBUG oslo_concurrency.lockutils [req-77ff7b41-f18f-42ac-8625-c2375117b229 req-2598e5af-360b-44c2-bf0a-1c93a6d74910 service nova] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1348.745044] env[70013]: DEBUG oslo_concurrency.lockutils [req-77ff7b41-f18f-42ac-8625-c2375117b229 req-2598e5af-360b-44c2-bf0a-1c93a6d74910 service nova] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1348.745044] env[70013]: DEBUG nova.compute.manager [req-77ff7b41-f18f-42ac-8625-c2375117b229 req-2598e5af-360b-44c2-bf0a-1c93a6d74910 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] No waiting events found dispatching network-vif-plugged-09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1348.745044] env[70013]: WARNING nova.compute.manager [req-77ff7b41-f18f-42ac-8625-c2375117b229 req-2598e5af-360b-44c2-bf0a-1c93a6d74910 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Received unexpected event network-vif-plugged-09fcfde5-581c-45f5-94e0-d8ed1b20d071 for instance with vm_state building and task_state spawning. [ 1349.038283] env[70013]: DEBUG nova.network.neutron [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Successfully updated port: 09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1349.543530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1349.543845] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1349.544134] env[70013]: DEBUG nova.network.neutron [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1350.080752] env[70013]: DEBUG nova.network.neutron [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1350.212571] env[70013]: DEBUG nova.network.neutron [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updating instance_info_cache with network_info: [{"id": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "address": "fa:16:3e:3e:6c:cf", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09fcfde5-58", "ovs_interfaceid": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1350.715181] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1350.715563] env[70013]: DEBUG nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Instance network_info: |[{"id": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "address": "fa:16:3e:3e:6c:cf", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09fcfde5-58", "ovs_interfaceid": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1350.716075] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:6c:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '09fcfde5-581c-45f5-94e0-d8ed1b20d071', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1350.723655] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating folder: Project (66804721098f48d69e216807291268d2). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1350.723950] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fcc2d9c2-c658-45cb-a14a-40b43b26d6d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.736084] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created folder: Project (66804721098f48d69e216807291268d2) in parent group-v836999. [ 1350.736301] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating folder: Instances. Parent ref: group-v837244. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1350.736557] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc0e8d0f-c5a0-496b-986d-9500cc04255c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.747212] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created folder: Instances in parent group-v837244. [ 1350.747483] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1350.747690] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1350.747917] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76f25ad4-77b7-40c8-9f49-9c371f67056b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.767977] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1350.767977] env[70013]: value = "task-4231658" [ 1350.767977] env[70013]: _type = "Task" [ 1350.767977] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1350.774918] env[70013]: DEBUG nova.compute.manager [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Received event network-changed-09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1350.774918] env[70013]: DEBUG nova.compute.manager [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Refreshing instance network info cache due to event network-changed-09fcfde5-581c-45f5-94e0-d8ed1b20d071. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1350.774918] env[70013]: DEBUG oslo_concurrency.lockutils [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] Acquiring lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1350.774918] env[70013]: DEBUG oslo_concurrency.lockutils [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] Acquired lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1350.774918] env[70013]: DEBUG nova.network.neutron [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Refreshing network info cache for port 09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1350.779350] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231658, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.282841] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231658, 'name': CreateVM_Task, 'duration_secs': 0.331443} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1351.282841] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1351.283679] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1351.283835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1351.284182] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1351.284458] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5301e4d4-cbae-4bb9-968b-6c1a60cd9c2a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.289498] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1351.289498] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e74dd-fe91-6dc8-2d13-fe447bdfccf1" [ 1351.289498] env[70013]: _type = "Task" [ 1351.289498] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1351.298825] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e74dd-fe91-6dc8-2d13-fe447bdfccf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.470397] env[70013]: DEBUG nova.network.neutron [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updated VIF entry in instance network info cache for port 09fcfde5-581c-45f5-94e0-d8ed1b20d071. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1351.470751] env[70013]: DEBUG nova.network.neutron [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updating instance_info_cache with network_info: [{"id": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "address": "fa:16:3e:3e:6c:cf", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09fcfde5-58", "ovs_interfaceid": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1351.800452] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]521e74dd-fe91-6dc8-2d13-fe447bdfccf1, 'name': SearchDatastore_Task, 'duration_secs': 0.010429} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1351.800854] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1351.801600] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1351.801902] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1351.802087] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1351.802278] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1351.802553] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7bdb3b93-ce51-48be-92b4-59e2f37fba8d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.812244] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1351.812422] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1351.813177] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb1f3308-ced9-4419-9006-6b5cd3733dc1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.819070] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1351.819070] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f4b097-3c84-0b70-110f-00ed1495343a" [ 1351.819070] env[70013]: _type = "Task" [ 1351.819070] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1351.827845] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f4b097-3c84-0b70-110f-00ed1495343a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.973917] env[70013]: DEBUG oslo_concurrency.lockutils [req-42edf384-74de-40a6-afa0-a1513afb35b4 req-76807a20-5fac-466d-a729-9e53c1215fda service nova] Releasing lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1352.329749] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f4b097-3c84-0b70-110f-00ed1495343a, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1352.330553] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5693d2e-a43d-45a6-9950-b6708b2facfe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.336330] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1352.336330] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5253b57e-f75f-c612-4628-106ea52d9592" [ 1352.336330] env[70013]: _type = "Task" [ 1352.336330] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.344115] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5253b57e-f75f-c612-4628-106ea52d9592, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1352.847847] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5253b57e-f75f-c612-4628-106ea52d9592, 'name': SearchDatastore_Task, 'duration_secs': 0.009681} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1352.848370] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1352.848490] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8d90b797-118d-45cb-9422-7137b7fadbe2/8d90b797-118d-45cb-9422-7137b7fadbe2.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1352.848687] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb06dbd6-ac48-4900-bc21-727459e94947 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.856214] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1352.856214] env[70013]: value = "task-4231659" [ 1352.856214] env[70013]: _type = "Task" [ 1352.856214] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.864357] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231659, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.368821] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231659, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445494} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1353.369059] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 8d90b797-118d-45cb-9422-7137b7fadbe2/8d90b797-118d-45cb-9422-7137b7fadbe2.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1353.369262] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1353.369535] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1dc261a1-a85b-491f-8cf2-92a71b9ab1c3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.378667] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1353.378667] env[70013]: value = "task-4231660" [ 1353.378667] env[70013]: _type = "Task" [ 1353.378667] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.387859] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.889184] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063103} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1353.889558] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1353.890208] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c5e7f8-cf90-4a93-a673-9352aa47a1be {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.911252] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 8d90b797-118d-45cb-9422-7137b7fadbe2/8d90b797-118d-45cb-9422-7137b7fadbe2.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1353.911491] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8265f36-a5e9-4511-9eef-4866144d8c3d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.933101] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1353.933101] env[70013]: value = "task-4231661" [ 1353.933101] env[70013]: _type = "Task" [ 1353.933101] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.940935] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.445351] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231661, 'name': ReconfigVM_Task, 'duration_secs': 0.283146} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.445600] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 8d90b797-118d-45cb-9422-7137b7fadbe2/8d90b797-118d-45cb-9422-7137b7fadbe2.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1354.446267] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e02e4bd-ade0-47bb-945f-559b32ac3b9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.453128] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1354.453128] env[70013]: value = "task-4231662" [ 1354.453128] env[70013]: _type = "Task" [ 1354.453128] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.461180] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231662, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.963408] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231662, 'name': Rename_Task, 'duration_secs': 0.142065} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.963852] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1354.963909] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-699a46a1-16fa-429f-8b11-609d4199617c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.970264] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1354.970264] env[70013]: value = "task-4231663" [ 1354.970264] env[70013]: _type = "Task" [ 1354.970264] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.977926] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231663, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.480977] env[70013]: DEBUG oslo_vmware.api [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231663, 'name': PowerOnVM_Task, 'duration_secs': 0.45022} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.481265] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1355.481471] env[70013]: INFO nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Took 6.94 seconds to spawn the instance on the hypervisor. [ 1355.481650] env[70013]: DEBUG nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1355.482493] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b3c839-0c23-4b8c-befe-2e2d1cb2b818 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.001443] env[70013]: INFO nova.compute.manager [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Took 11.71 seconds to build instance. [ 1356.503050] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d69288dd-7487-4a3d-ba74-2d44c15e0443 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.218s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1356.513491] env[70013]: DEBUG nova.compute.manager [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Received event network-changed-09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1356.513698] env[70013]: DEBUG nova.compute.manager [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Refreshing instance network info cache due to event network-changed-09fcfde5-581c-45f5-94e0-d8ed1b20d071. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1356.513922] env[70013]: DEBUG oslo_concurrency.lockutils [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] Acquiring lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1356.514088] env[70013]: DEBUG oslo_concurrency.lockutils [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] Acquired lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1356.514260] env[70013]: DEBUG nova.network.neutron [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Refreshing network info cache for port 09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1357.225730] env[70013]: DEBUG nova.network.neutron [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updated VIF entry in instance network info cache for port 09fcfde5-581c-45f5-94e0-d8ed1b20d071. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1357.226116] env[70013]: DEBUG nova.network.neutron [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updating instance_info_cache with network_info: [{"id": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "address": "fa:16:3e:3e:6c:cf", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09fcfde5-58", "ovs_interfaceid": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1357.728948] env[70013]: DEBUG oslo_concurrency.lockutils [req-7f583d3c-f22f-47f5-89db-9eef7e80121d req-8ae616f2-600b-498e-82ad-b7469afb5c1d service nova] Releasing lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1361.340162] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1362.340450] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1362.843977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1362.844213] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1362.844391] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1362.844546] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1362.845587] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c25360b-4c0d-4864-8fa5-aa253be165b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.854488] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756e532a-8418-48d6-a013-8a24663f0ab1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.870266] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e12970-7c3a-4a4f-8681-517e2e8980da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.877170] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a51035-587d-4b25-b98e-502416a76910 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.906142] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180422MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1362.906350] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1362.906532] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1363.941701] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.941998] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.941998] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.942103] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.942221] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.942391] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.942562] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.942683] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.942854] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8b74079f-e9d3-4212-8b7c-31f59e75ca13 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.942994] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8d90b797-118d-45cb-9422-7137b7fadbe2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1363.943138] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1363.943278] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1363.959036] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1363.972243] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1363.972464] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1363.984408] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1364.003142] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1364.117249] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6727739c-fb22-4531-9b3e-045d73d3bf87 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.124852] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5543cf10-3fff-4d13-a632-a9f40fc389b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.155845] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74b34ab-80f7-4620-9508-7f933fcf2438 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.163609] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a20aa3b-e15e-4006-9748-c193bf3e0839 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.177169] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1364.680124] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1365.185039] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1365.185553] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.279s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1366.185630] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.186077] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.186148] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.186393] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.186609] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.186793] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1367.335588] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1370.340281] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.486578] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1374.486926] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1374.990663] env[70013]: DEBUG nova.compute.utils [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1375.493655] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1376.559816] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1376.560241] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1376.560365] env[70013]: INFO nova.compute.manager [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Attaching volume d2d92094-e0c4-41f8-9c4a-e68cac80f315 to /dev/sdb [ 1376.591960] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59734c84-a4a9-4349-8cd6-bbe43a55904a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.599503] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0725b2c8-df54-42cf-8298-53b7f15828d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.613872] env[70013]: DEBUG nova.virt.block_device [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating existing volume attachment record: 0c78c7ca-aaf4-402c-9bb0-2ad0f2a23e6b {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1381.159126] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1381.159432] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837248', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'name': 'volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8b74079f-e9d3-4212-8b7c-31f59e75ca13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'serial': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1381.160355] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc55dab-4813-4d34-9191-2ecc0ea50d86 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.178286] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0454455-4ace-453a-a9fe-5ae7ee7162ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.204047] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315/volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1381.204368] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4eac5d46-2401-4415-ab13-fba30c4d6f6a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.223163] env[70013]: DEBUG oslo_vmware.api [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1381.223163] env[70013]: value = "task-4231668" [ 1381.223163] env[70013]: _type = "Task" [ 1381.223163] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1381.232100] env[70013]: DEBUG oslo_vmware.api [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231668, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1381.734186] env[70013]: DEBUG oslo_vmware.api [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231668, 'name': ReconfigVM_Task, 'duration_secs': 0.351319} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.734466] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315/volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1381.739199] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-221b727f-00c7-4525-a779-13c324ade601 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.754748] env[70013]: DEBUG oslo_vmware.api [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1381.754748] env[70013]: value = "task-4231669" [ 1381.754748] env[70013]: _type = "Task" [ 1381.754748] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1381.763780] env[70013]: DEBUG oslo_vmware.api [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231669, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.264932] env[70013]: DEBUG oslo_vmware.api [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231669, 'name': ReconfigVM_Task, 'duration_secs': 0.138375} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1382.265306] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837248', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'name': 'volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8b74079f-e9d3-4212-8b7c-31f59e75ca13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'serial': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1383.313068] env[70013]: DEBUG nova.objects.instance [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1383.820011] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6cfffd53-c821-4749-9c7b-d20b8ea2da6f tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.259s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1383.904041] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1383.904309] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1383.904519] env[70013]: DEBUG nova.compute.manager [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1383.905492] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050a84f6-a707-4fd6-862e-443fdcc48f1b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.912613] env[70013]: DEBUG nova.compute.manager [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1383.913204] env[70013]: DEBUG nova.objects.instance [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1384.922427] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1384.922829] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f0af90c-86bb-48ba-a224-96b3d84d970b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.930688] env[70013]: DEBUG oslo_vmware.api [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1384.930688] env[70013]: value = "task-4231670" [ 1384.930688] env[70013]: _type = "Task" [ 1384.930688] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1384.939993] env[70013]: DEBUG oslo_vmware.api [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.440679] env[70013]: DEBUG oslo_vmware.api [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231670, 'name': PowerOffVM_Task, 'duration_secs': 0.186796} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1385.440937] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1385.441162] env[70013]: DEBUG nova.compute.manager [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1385.441953] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082d135f-0bd8-4888-a2ec-92e5fc6cfb8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.954999] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e6be94e8-9eef-47c2-8dd5-fe691eb01cbd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.051s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1386.280274] env[70013]: DEBUG nova.objects.instance [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1386.786067] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1386.786290] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1386.786457] env[70013]: DEBUG nova.network.neutron [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1386.786636] env[70013]: DEBUG nova.objects.instance [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'info_cache' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1387.290519] env[70013]: DEBUG nova.objects.base [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Object Instance<8b74079f-e9d3-4212-8b7c-31f59e75ca13> lazy-loaded attributes: flavor,info_cache {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1388.016228] env[70013]: DEBUG nova.network.neutron [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating instance_info_cache with network_info: [{"id": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "address": "fa:16:3e:f6:03:32", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0641d8ef-98", "ovs_interfaceid": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1388.519433] env[70013]: DEBUG oslo_concurrency.lockutils [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1389.526418] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1389.526884] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ba3dba3-a9d4-4051-bd2f-abf2c9855052 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.534429] env[70013]: DEBUG oslo_vmware.api [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1389.534429] env[70013]: value = "task-4231671" [ 1389.534429] env[70013]: _type = "Task" [ 1389.534429] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.543307] env[70013]: DEBUG oslo_vmware.api [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.045656] env[70013]: DEBUG oslo_vmware.api [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231671, 'name': PowerOnVM_Task, 'duration_secs': 0.449958} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.045917] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1390.046145] env[70013]: DEBUG nova.compute.manager [None req-e77f9d13-d26d-4e04-8a6b-1c31b1052fac tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1390.047023] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17d9325-4c08-4ff7-8130-677918613e94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.796652] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1394.797085] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1395.299572] env[70013]: DEBUG nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1395.819054] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1395.819436] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1395.820856] env[70013]: INFO nova.compute.claims [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1396.972937] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0e7ada-58a7-4535-80a0-b7952f356ad8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.981653] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6a01b9-7173-412e-b080-822b95c70e37 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.012135] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d2c445-ec26-4d11-9c12-d27ec6a0851f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.020425] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9117833-1019-4b5f-9072-7855d6347715 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.034776] env[70013]: DEBUG nova.compute.provider_tree [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1397.538132] env[70013]: DEBUG nova.scheduler.client.report [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1398.044061] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1398.044061] env[70013]: DEBUG nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1398.549737] env[70013]: DEBUG nova.compute.utils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1398.551601] env[70013]: DEBUG nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1398.551813] env[70013]: DEBUG nova.network.neutron [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1398.603171] env[70013]: DEBUG nova.policy [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ade87067e834404adc174783ff7b03d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66804721098f48d69e216807291268d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1398.887909] env[70013]: DEBUG nova.network.neutron [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Successfully created port: 94489867-3216-44ac-bd4b-224974ef5c89 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1399.055208] env[70013]: DEBUG nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1400.067415] env[70013]: DEBUG nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1400.096209] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1400.096477] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1400.096632] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1400.096817] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1400.096960] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1400.097129] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1400.097343] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1400.097961] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1400.097961] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1400.097961] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1400.098146] env[70013]: DEBUG nova.virt.hardware [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1400.098902] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4929cb1-7b52-4a71-b25a-cffb70ec6140 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.107416] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bea5fd2-0ece-403d-8dfd-b803d744ba7d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.266782] env[70013]: DEBUG nova.compute.manager [req-dc67e1fc-dc33-4f49-b3a8-e2ca228b6568 req-3592990e-fb80-417c-9c99-08f6ebf1acb3 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Received event network-vif-plugged-94489867-3216-44ac-bd4b-224974ef5c89 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1400.267010] env[70013]: DEBUG oslo_concurrency.lockutils [req-dc67e1fc-dc33-4f49-b3a8-e2ca228b6568 req-3592990e-fb80-417c-9c99-08f6ebf1acb3 service nova] Acquiring lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1400.267230] env[70013]: DEBUG oslo_concurrency.lockutils [req-dc67e1fc-dc33-4f49-b3a8-e2ca228b6568 req-3592990e-fb80-417c-9c99-08f6ebf1acb3 service nova] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1400.267398] env[70013]: DEBUG oslo_concurrency.lockutils [req-dc67e1fc-dc33-4f49-b3a8-e2ca228b6568 req-3592990e-fb80-417c-9c99-08f6ebf1acb3 service nova] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1400.267562] env[70013]: DEBUG nova.compute.manager [req-dc67e1fc-dc33-4f49-b3a8-e2ca228b6568 req-3592990e-fb80-417c-9c99-08f6ebf1acb3 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] No waiting events found dispatching network-vif-plugged-94489867-3216-44ac-bd4b-224974ef5c89 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1400.267774] env[70013]: WARNING nova.compute.manager [req-dc67e1fc-dc33-4f49-b3a8-e2ca228b6568 req-3592990e-fb80-417c-9c99-08f6ebf1acb3 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Received unexpected event network-vif-plugged-94489867-3216-44ac-bd4b-224974ef5c89 for instance with vm_state building and task_state spawning. [ 1400.355495] env[70013]: DEBUG nova.network.neutron [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Successfully updated port: 94489867-3216-44ac-bd4b-224974ef5c89 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1400.858156] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-2f7b0d3d-4e6f-458a-81af-73a525f33df3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1400.858345] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-2f7b0d3d-4e6f-458a-81af-73a525f33df3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1400.858604] env[70013]: DEBUG nova.network.neutron [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1401.392720] env[70013]: DEBUG nova.network.neutron [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1401.527166] env[70013]: DEBUG nova.network.neutron [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Updating instance_info_cache with network_info: [{"id": "94489867-3216-44ac-bd4b-224974ef5c89", "address": "fa:16:3e:ec:e5:22", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94489867-32", "ovs_interfaceid": "94489867-3216-44ac-bd4b-224974ef5c89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1402.030883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-2f7b0d3d-4e6f-458a-81af-73a525f33df3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1402.031398] env[70013]: DEBUG nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Instance network_info: |[{"id": "94489867-3216-44ac-bd4b-224974ef5c89", "address": "fa:16:3e:ec:e5:22", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94489867-32", "ovs_interfaceid": "94489867-3216-44ac-bd4b-224974ef5c89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1402.031991] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:e5:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94489867-3216-44ac-bd4b-224974ef5c89', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1402.042791] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1402.043103] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1402.043429] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-88768cdd-7854-4ec8-bd55-1883c6f6335e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.071242] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1402.071242] env[70013]: value = "task-4231672" [ 1402.071242] env[70013]: _type = "Task" [ 1402.071242] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.079488] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231672, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.296916] env[70013]: DEBUG nova.compute.manager [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Received event network-changed-94489867-3216-44ac-bd4b-224974ef5c89 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1402.297254] env[70013]: DEBUG nova.compute.manager [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Refreshing instance network info cache due to event network-changed-94489867-3216-44ac-bd4b-224974ef5c89. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1402.297393] env[70013]: DEBUG oslo_concurrency.lockutils [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] Acquiring lock "refresh_cache-2f7b0d3d-4e6f-458a-81af-73a525f33df3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.297547] env[70013]: DEBUG oslo_concurrency.lockutils [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] Acquired lock "refresh_cache-2f7b0d3d-4e6f-458a-81af-73a525f33df3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1402.297712] env[70013]: DEBUG nova.network.neutron [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Refreshing network info cache for port 94489867-3216-44ac-bd4b-224974ef5c89 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1402.580841] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231672, 'name': CreateVM_Task, 'duration_secs': 0.327257} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1402.581220] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1402.581677] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.581863] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1402.582211] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1402.582473] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-564417f1-bc8e-469e-a9a6-d0833db42060 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.587637] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1402.587637] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5288fbcc-75b2-b34a-9297-e47a21ea2c17" [ 1402.587637] env[70013]: _type = "Task" [ 1402.587637] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.597189] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5288fbcc-75b2-b34a-9297-e47a21ea2c17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.995143] env[70013]: DEBUG nova.network.neutron [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Updated VIF entry in instance network info cache for port 94489867-3216-44ac-bd4b-224974ef5c89. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1402.995572] env[70013]: DEBUG nova.network.neutron [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Updating instance_info_cache with network_info: [{"id": "94489867-3216-44ac-bd4b-224974ef5c89", "address": "fa:16:3e:ec:e5:22", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94489867-32", "ovs_interfaceid": "94489867-3216-44ac-bd4b-224974ef5c89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1403.098573] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5288fbcc-75b2-b34a-9297-e47a21ea2c17, 'name': SearchDatastore_Task, 'duration_secs': 0.012605} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1403.098896] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1403.099155] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1403.099405] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1403.099548] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1403.099765] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1403.100058] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa49aa79-de92-4053-bbdb-1ce24b0cb29a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.109674] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1403.109876] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1403.110620] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fd320e0-f5dd-4686-8ee8-28dc7ff569c2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.116556] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1403.116556] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]527b9a5f-aabe-7b5e-9308-7e94d913e49c" [ 1403.116556] env[70013]: _type = "Task" [ 1403.116556] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.124497] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527b9a5f-aabe-7b5e-9308-7e94d913e49c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.498550] env[70013]: DEBUG oslo_concurrency.lockutils [req-416d0f41-3f5f-4f98-bbd4-aaa79ed92614 req-4e1a2cad-cd46-4369-9773-5cd869d6ddd7 service nova] Releasing lock "refresh_cache-2f7b0d3d-4e6f-458a-81af-73a525f33df3" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1403.627864] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]527b9a5f-aabe-7b5e-9308-7e94d913e49c, 'name': SearchDatastore_Task, 'duration_secs': 0.009992} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1403.628732] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-843b9edf-0569-4dd9-9bac-6ceb4c2088d4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.635399] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1403.635399] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3431e-5b93-fbd7-a834-a68cebd7f6be" [ 1403.635399] env[70013]: _type = "Task" [ 1403.635399] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.643498] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3431e-5b93-fbd7-a834-a68cebd7f6be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.145660] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3431e-5b93-fbd7-a834-a68cebd7f6be, 'name': SearchDatastore_Task, 'duration_secs': 0.010397} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.145904] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1404.146135] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1404.146432] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-579ffb79-45a0-4d05-9cdb-abb1647dfb79 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.153976] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1404.153976] env[70013]: value = "task-4231673" [ 1404.153976] env[70013]: _type = "Task" [ 1404.153976] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.162140] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.663877] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473828} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.664281] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1404.664422] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1404.664620] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c88f1db5-64f0-4c6e-848b-17988420230d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.671606] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1404.671606] env[70013]: value = "task-4231674" [ 1404.671606] env[70013]: _type = "Task" [ 1404.671606] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.681026] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.182154] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068131} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.182418] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1405.183261] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbee16c3-c63b-456b-9121-d20cc56f8d9c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.206609] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1405.206981] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0896d66d-dbeb-4670-8e6b-7b1451a7ff75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.227169] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1405.227169] env[70013]: value = "task-4231675" [ 1405.227169] env[70013]: _type = "Task" [ 1405.227169] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.235756] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231675, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.739016] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231675, 'name': ReconfigVM_Task} progress is 14%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.238642] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231675, 'name': ReconfigVM_Task, 'duration_secs': 0.698115} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.238922] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1406.239599] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a2eac6c-408f-4aa6-86a0-e018040339a6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.246090] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1406.246090] env[70013]: value = "task-4231676" [ 1406.246090] env[70013]: _type = "Task" [ 1406.246090] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.255353] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231676, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.757881] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231676, 'name': Rename_Task, 'duration_secs': 0.146993} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.758281] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1406.758433] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b6c5636-0e1d-47e7-a65b-57f8a4696062 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.766217] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1406.766217] env[70013]: value = "task-4231677" [ 1406.766217] env[70013]: _type = "Task" [ 1406.766217] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.774274] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.277022] env[70013]: DEBUG oslo_vmware.api [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231677, 'name': PowerOnVM_Task, 'duration_secs': 0.441722} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.277022] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1407.277022] env[70013]: INFO nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Took 7.21 seconds to spawn the instance on the hypervisor. [ 1407.277200] env[70013]: DEBUG nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1407.277891] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b323daf4-5654-46c4-8d59-65339fd61f31 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.795963] env[70013]: INFO nova.compute.manager [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Took 11.99 seconds to build instance. [ 1408.297971] env[70013]: DEBUG oslo_concurrency.lockutils [None req-77c0b80f-e556-41c2-8a90-1b5ee48c173c tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.501s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1409.089084] env[70013]: DEBUG oslo_concurrency.lockutils [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1409.089084] env[70013]: DEBUG oslo_concurrency.lockutils [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1409.089084] env[70013]: DEBUG nova.compute.manager [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1409.089981] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37158066-3390-419b-813f-cc30a4f21836 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.097046] env[70013]: DEBUG nova.compute.manager [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1409.097618] env[70013]: DEBUG nova.objects.instance [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'flavor' on Instance uuid 2f7b0d3d-4e6f-458a-81af-73a525f33df3 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1410.104808] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1410.105179] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-561673fd-c670-4441-8dfa-e810b8b4fc66 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.113051] env[70013]: DEBUG oslo_vmware.api [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1410.113051] env[70013]: value = "task-4231678" [ 1410.113051] env[70013]: _type = "Task" [ 1410.113051] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1410.121565] env[70013]: DEBUG oslo_vmware.api [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.624139] env[70013]: DEBUG oslo_vmware.api [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231678, 'name': PowerOffVM_Task, 'duration_secs': 0.190833} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1410.624139] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1410.624139] env[70013]: DEBUG nova.compute.manager [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1410.624852] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc87f75-16a1-4ce2-87c1-501c25841492 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.137923] env[70013]: DEBUG oslo_concurrency.lockutils [None req-57f026b0-6125-407b-9e04-7e4d5cd975a7 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.049s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1411.990440] env[70013]: INFO nova.compute.manager [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Rebuilding instance [ 1412.035244] env[70013]: DEBUG nova.compute.manager [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1412.036185] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fb8ef6-9f9a-4d4b-9788-0db285d01d05 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.051427] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1413.051869] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6db84b36-ec44-4767-b688-44dc38a2ecf8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.060832] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1413.060832] env[70013]: value = "task-4231679" [ 1413.060832] env[70013]: _type = "Task" [ 1413.060832] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1413.070772] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1413.071032] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1413.071782] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a835dfec-364c-4a23-a5f5-b384b6d19394 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.078922] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1413.079189] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ab1d91e-0453-4f7f-b57d-2b69e2af0949 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.154127] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1413.154372] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1413.154541] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleting the datastore file [datastore2] 2f7b0d3d-4e6f-458a-81af-73a525f33df3 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1413.154820] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d88ace3c-3861-49d9-911b-4360f7396951 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.161770] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1413.161770] env[70013]: value = "task-4231681" [ 1413.161770] env[70013]: _type = "Task" [ 1413.161770] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1413.170738] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231681, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1413.671944] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231681, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134007} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1413.672231] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1413.672416] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1413.672587] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1414.703351] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1414.703651] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1414.703738] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1414.703919] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1414.704076] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1414.704225] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1414.704428] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1414.704587] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1414.704752] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1414.704911] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1414.705094] env[70013]: DEBUG nova.virt.hardware [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1414.705978] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e42b57-27c9-4924-904a-6399e32816f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.714169] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821255f6-5e47-4d21-b381-e86c6f65e177 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.728376] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:e5:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94489867-3216-44ac-bd4b-224974ef5c89', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1414.735813] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1414.736068] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1414.736293] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f53af93-269b-47d3-91ed-c14fee1b28cf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.756737] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1414.756737] env[70013]: value = "task-4231682" [ 1414.756737] env[70013]: _type = "Task" [ 1414.756737] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1414.765725] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231682, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1415.266300] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231682, 'name': CreateVM_Task, 'duration_secs': 0.333896} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1415.266534] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1415.267178] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1415.267354] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1415.267708] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1415.267957] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c0076f5-0f27-4633-b7e0-edb76b0bff0e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.272578] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1415.272578] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]523fe46d-f133-e906-1eff-486cf86fc314" [ 1415.272578] env[70013]: _type = "Task" [ 1415.272578] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1415.280323] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523fe46d-f133-e906-1eff-486cf86fc314, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1415.783729] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]523fe46d-f133-e906-1eff-486cf86fc314, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1415.784123] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1415.784235] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1415.784470] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1415.784617] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1415.784792] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1415.785075] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a71f27eb-32d5-44e3-ba2b-a3d8cc861e15 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.794741] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1415.794930] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1415.795738] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e30a2636-dd97-4421-bb6e-0db4852454a0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.801779] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1415.801779] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]520624ef-7118-e13b-5d75-69a5692f4a0f" [ 1415.801779] env[70013]: _type = "Task" [ 1415.801779] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1415.810023] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520624ef-7118-e13b-5d75-69a5692f4a0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.313166] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]520624ef-7118-e13b-5d75-69a5692f4a0f, 'name': SearchDatastore_Task, 'duration_secs': 0.010065} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.313975] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9bcf87e-e1c1-431e-977a-6877242cc8d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.319402] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1416.319402] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e296f7-7bde-2fff-4227-42506f3f2510" [ 1416.319402] env[70013]: _type = "Task" [ 1416.319402] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.327604] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e296f7-7bde-2fff-4227-42506f3f2510, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.830658] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e296f7-7bde-2fff-4227-42506f3f2510, 'name': SearchDatastore_Task, 'duration_secs': 0.010343} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.831075] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1416.831202] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1416.831458] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d9cd008-fd24-449c-b55e-d914c98c2ead {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.839308] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1416.839308] env[70013]: value = "task-4231683" [ 1416.839308] env[70013]: _type = "Task" [ 1416.839308] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.848567] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231683, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.348850] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231683, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480875} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.349115] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1417.349322] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1417.349573] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9913717-21f8-404c-973d-d15486961bab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.356036] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1417.356036] env[70013]: value = "task-4231684" [ 1417.356036] env[70013]: _type = "Task" [ 1417.356036] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.365029] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231684, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.866291] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231684, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06168} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.866691] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1417.867437] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc94de68-5d32-43cd-98f5-06250c178c6e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.889465] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1417.889750] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04545753-7f94-4647-b04a-31bcbad0240d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.909858] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1417.909858] env[70013]: value = "task-4231685" [ 1417.909858] env[70013]: _type = "Task" [ 1417.909858] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.918064] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231685, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.421590] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231685, 'name': ReconfigVM_Task, 'duration_secs': 0.282031} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.421879] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 2f7b0d3d-4e6f-458a-81af-73a525f33df3/2f7b0d3d-4e6f-458a-81af-73a525f33df3.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1418.422566] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-933cc468-32f9-4538-90ce-64f58eb1688c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.429602] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1418.429602] env[70013]: value = "task-4231686" [ 1418.429602] env[70013]: _type = "Task" [ 1418.429602] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.437832] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231686, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.939615] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231686, 'name': Rename_Task, 'duration_secs': 0.151297} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.939981] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1418.940144] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08d749ec-928d-4d57-97ab-a424a0447bdc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.947586] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1418.947586] env[70013]: value = "task-4231687" [ 1418.947586] env[70013]: _type = "Task" [ 1418.947586] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.956530] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231687, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1419.457961] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231687, 'name': PowerOnVM_Task, 'duration_secs': 0.446803} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1419.458254] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1419.458458] env[70013]: DEBUG nova.compute.manager [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1419.459272] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e46477-8439-4643-b37b-f9bdf05b3334 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.970846] env[70013]: INFO nova.compute.manager [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] bringing vm to original state: 'stopped' [ 1420.979087] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1420.979478] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1420.979678] env[70013]: DEBUG nova.compute.manager [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1420.980682] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e155131b-cc9f-4ec8-b890-0580cd53465e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.988771] env[70013]: DEBUG nova.compute.manager [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1421.339849] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1421.493975] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1421.494301] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f76c020f-8136-4c96-90a4-bbb7153fff98 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.501373] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1421.501373] env[70013]: value = "task-4231688" [ 1421.501373] env[70013]: _type = "Task" [ 1421.501373] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.510355] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1422.011713] env[70013]: DEBUG oslo_vmware.api [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231688, 'name': PowerOffVM_Task, 'duration_secs': 0.20992} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1422.012112] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1422.012112] env[70013]: DEBUG nova.compute.manager [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1422.012880] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766711a7-a738-461c-9dca-765204d85ebf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.339715] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1422.526369] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.547s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1422.843783] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1422.844031] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1422.844201] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1422.844356] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1422.845288] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaa8c29-4b26-479b-ae8e-978eeb1a6205 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.853795] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45be3184-b42d-458c-b86a-148974ded681 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.868786] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adad29ca-8e00-4817-bf07-4382d6685f2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.876199] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2176ccd-222b-4211-a150-eb431011b28e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.908076] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180161MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1422.908076] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1422.908076] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1423.020416] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1423.020800] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1423.020920] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1423.021127] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1423.021300] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1423.023485] env[70013]: INFO nova.compute.manager [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Terminating instance [ 1423.035829] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1423.527125] env[70013]: DEBUG nova.compute.manager [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1423.527404] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1423.528275] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7316820-79de-4393-9c26-462fe75fcb8e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.537255] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1423.537486] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f14af04-9468-4fbf-9f26-6dfd49e1668b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.606544] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1423.606812] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1423.607086] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleting the datastore file [datastore1] 2f7b0d3d-4e6f-458a-81af-73a525f33df3 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1423.607385] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ae6c607-dc7d-4db2-ba87-c4dbb2c7e5e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.614568] env[70013]: DEBUG oslo_vmware.api [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1423.614568] env[70013]: value = "task-4231690" [ 1423.614568] env[70013]: _type = "Task" [ 1423.614568] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.624019] env[70013]: DEBUG oslo_vmware.api [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1423.943794] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.943967] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944092] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944214] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944328] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944444] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944556] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944667] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944776] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8b74079f-e9d3-4212-8b7c-31f59e75ca13 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944886] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8d90b797-118d-45cb-9422-7137b7fadbe2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.944995] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2f7b0d3d-4e6f-458a-81af-73a525f33df3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1423.945309] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1423.945472] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1424.092086] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb533d4a-90e4-4836-bda8-31c535d12bfb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.100749] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3d8f2a-f081-4e35-8ced-a0fa66b74f03 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.134652] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80c97aa-c81f-44ee-8529-1a3da08ea7bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.145648] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e47ec8-b510-4e2f-a767-2d9a40dbefa7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.149462] env[70013]: DEBUG oslo_vmware.api [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161261} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.149703] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1424.149888] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1424.150072] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1424.150247] env[70013]: INFO nova.compute.manager [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1424.150492] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1424.151089] env[70013]: DEBUG nova.compute.manager [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1424.151189] env[70013]: DEBUG nova.network.neutron [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1424.161188] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1424.434779] env[70013]: DEBUG nova.compute.manager [req-80a9ee42-508d-4e64-812c-3a03dd4b127b req-f711003f-b12b-4ddf-81ce-35d76423f625 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Received event network-vif-deleted-94489867-3216-44ac-bd4b-224974ef5c89 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1424.435026] env[70013]: INFO nova.compute.manager [req-80a9ee42-508d-4e64-812c-3a03dd4b127b req-f711003f-b12b-4ddf-81ce-35d76423f625 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Neutron deleted interface 94489867-3216-44ac-bd4b-224974ef5c89; detaching it from the instance and deleting it from the info cache [ 1424.435268] env[70013]: DEBUG nova.network.neutron [req-80a9ee42-508d-4e64-812c-3a03dd4b127b req-f711003f-b12b-4ddf-81ce-35d76423f625 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1424.666369] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1424.911191] env[70013]: DEBUG nova.network.neutron [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1424.937651] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4cc911f0-29c9-45ad-8c9d-fb2427e6d99f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.947535] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed9f881-d06d-44c9-8822-72a0f2b5942c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.974059] env[70013]: DEBUG nova.compute.manager [req-80a9ee42-508d-4e64-812c-3a03dd4b127b req-f711003f-b12b-4ddf-81ce-35d76423f625 service nova] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Detach interface failed, port_id=94489867-3216-44ac-bd4b-224974ef5c89, reason: Instance 2f7b0d3d-4e6f-458a-81af-73a525f33df3 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1425.172568] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1425.172989] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.265s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1425.173061] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.137s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1425.173244] env[70013]: DEBUG nova.objects.instance [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1425.414157] env[70013]: INFO nova.compute.manager [-] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Took 1.26 seconds to deallocate network for instance. [ 1425.921522] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1426.182789] env[70013]: DEBUG oslo_concurrency.lockutils [None req-ee0c22dc-6091-40e6-b466-b1cbc2e4e498 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1426.184085] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.263s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1426.184323] env[70013]: DEBUG nova.objects.instance [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'resources' on Instance uuid 2f7b0d3d-4e6f-458a-81af-73a525f33df3 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1426.428184] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1426.428480] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1426.833146] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15330bc8-9283-4695-9403-96caa3dc5459 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.841333] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a68cdb8-ebbd-4508-9ecc-bdab29082dbd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.872955] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d54b36-73b3-4339-a29e-d4736ef93d23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.880931] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36fc113-d56d-4cf1-b90c-42fc6bd53f30 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.894816] env[70013]: DEBUG nova.compute.provider_tree [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1426.931711] env[70013]: INFO nova.compute.manager [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Detaching volume d2d92094-e0c4-41f8-9c4a-e68cac80f315 [ 1426.964930] env[70013]: INFO nova.virt.block_device [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Attempting to driver detach volume d2d92094-e0c4-41f8-9c4a-e68cac80f315 from mountpoint /dev/sdb [ 1426.965183] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1426.965374] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837248', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'name': 'volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8b74079f-e9d3-4212-8b7c-31f59e75ca13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'serial': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1426.966457] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc43f33-cbca-43b6-9645-a11bb9820bfb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.989862] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19cb407-a775-4498-8c2b-a2d6a3532be4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.998011] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59b662f-3324-4c7d-afa3-be7e78977aae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.020051] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0fcade-5026-4380-b70b-21e4d10b2e23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.038628] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] The volume has not been displaced from its original location: [datastore2] volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315/volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1427.043889] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1427.044242] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c619af71-e309-440b-a82c-a1a94778fef4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.063153] env[70013]: DEBUG oslo_vmware.api [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1427.063153] env[70013]: value = "task-4231691" [ 1427.063153] env[70013]: _type = "Task" [ 1427.063153] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1427.074859] env[70013]: DEBUG oslo_vmware.api [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1427.398102] env[70013]: DEBUG nova.scheduler.client.report [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1427.573531] env[70013]: DEBUG oslo_vmware.api [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231691, 'name': ReconfigVM_Task, 'duration_secs': 0.246609} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1427.573748] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1427.578467] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c8b50d4-fa01-4697-838d-be6668a426a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.593900] env[70013]: DEBUG oslo_vmware.api [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1427.593900] env[70013]: value = "task-4231692" [ 1427.593900] env[70013]: _type = "Task" [ 1427.593900] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1427.602459] env[70013]: DEBUG oslo_vmware.api [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231692, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1427.903593] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1427.927274] env[70013]: INFO nova.scheduler.client.report [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted allocations for instance 2f7b0d3d-4e6f-458a-81af-73a525f33df3 [ 1428.104527] env[70013]: DEBUG oslo_vmware.api [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231692, 'name': ReconfigVM_Task, 'duration_secs': 0.142304} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1428.104833] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837248', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'name': 'volume-d2d92094-e0c4-41f8-9c4a-e68cac80f315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8b74079f-e9d3-4212-8b7c-31f59e75ca13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315', 'serial': 'd2d92094-e0c4-41f8-9c4a-e68cac80f315'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1428.169693] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1428.169878] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1428.170081] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1428.170244] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1428.170402] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1428.170541] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1428.435018] env[70013]: DEBUG oslo_concurrency.lockutils [None req-be793570-6671-4b6c-95fd-6bad0d41477d tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "2f7b0d3d-4e6f-458a-81af-73a525f33df3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.414s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1428.647459] env[70013]: DEBUG nova.objects.instance [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1429.596507] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1429.654406] env[70013]: DEBUG oslo_concurrency.lockutils [None req-aa2fdd06-2b1d-4a12-81d8-98f255032858 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.226s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1429.655628] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.059s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1429.655835] env[70013]: DEBUG nova.compute.manager [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1429.656900] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5878b2-0f4b-4522-b5ea-0ec0d5c618aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.664070] env[70013]: DEBUG nova.compute.manager [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=70013) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1429.664646] env[70013]: DEBUG nova.objects.instance [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1430.671627] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1430.672049] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f462c1f-0810-48ef-9eb4-6b4ba3c219e8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.679640] env[70013]: DEBUG oslo_vmware.api [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1430.679640] env[70013]: value = "task-4231693" [ 1430.679640] env[70013]: _type = "Task" [ 1430.679640] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.688013] env[70013]: DEBUG oslo_vmware.api [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.856217] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1430.856526] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1431.189669] env[70013]: DEBUG oslo_vmware.api [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231693, 'name': PowerOffVM_Task, 'duration_secs': 0.189256} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.189917] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1431.190114] env[70013]: DEBUG nova.compute.manager [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1431.190884] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b64f4f3-2f60-4fab-b856-c573f1501c57 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.340052] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1431.358825] env[70013]: DEBUG nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1431.703417] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2718179f-2d9c-4d40-ab96-5394e10d10fd tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.048s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1431.881569] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1431.881849] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1431.883441] env[70013]: INFO nova.compute.claims [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1432.001408] env[70013]: DEBUG nova.objects.instance [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1432.506247] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1432.506435] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1432.506632] env[70013]: DEBUG nova.network.neutron [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1432.506830] env[70013]: DEBUG nova.objects.instance [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'info_cache' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1433.010182] env[70013]: DEBUG nova.objects.base [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Object Instance<8b74079f-e9d3-4212-8b7c-31f59e75ca13> lazy-loaded attributes: flavor,info_cache {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1433.032608] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e453fc-7a47-46d0-b8d2-5319ea1e9291 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.040767] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f342a268-96cf-4d54-a36e-0cd5ff89ffb4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.070998] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e252fdfc-eb5c-4afe-b6e4-c442cbabf6ae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.078656] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed21d73-49bb-4dba-b299-a3d2fd4c7ea5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.092018] env[70013]: DEBUG nova.compute.provider_tree [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1433.595165] env[70013]: DEBUG nova.scheduler.client.report [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1433.731930] env[70013]: DEBUG nova.network.neutron [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating instance_info_cache with network_info: [{"id": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "address": "fa:16:3e:f6:03:32", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0641d8ef-98", "ovs_interfaceid": "0641d8ef-98b5-48f3-a1a8-6bdafa0154d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1434.100503] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1434.101143] env[70013]: DEBUG nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1434.234939] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "refresh_cache-8b74079f-e9d3-4212-8b7c-31f59e75ca13" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1434.607073] env[70013]: DEBUG nova.compute.utils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1434.608597] env[70013]: DEBUG nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1434.608788] env[70013]: DEBUG nova.network.neutron [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1434.648551] env[70013]: DEBUG nova.policy [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ade87067e834404adc174783ff7b03d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66804721098f48d69e216807291268d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1434.917603] env[70013]: DEBUG nova.network.neutron [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Successfully created port: 38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1435.112308] env[70013]: DEBUG nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1435.241225] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1435.241541] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52db0c07-37d6-483c-946f-e00f3bbf7f16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.249073] env[70013]: DEBUG oslo_vmware.api [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1435.249073] env[70013]: value = "task-4231694" [ 1435.249073] env[70013]: _type = "Task" [ 1435.249073] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1435.257226] env[70013]: DEBUG oslo_vmware.api [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231694, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1435.761017] env[70013]: DEBUG oslo_vmware.api [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231694, 'name': PowerOnVM_Task, 'duration_secs': 0.395029} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1435.761308] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1435.761506] env[70013]: DEBUG nova.compute.manager [None req-b6f3d74f-701b-4c18-b050-a1cb23410aae tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1435.762274] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf44f933-46b6-45c0-8b9c-59ca6b82722c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.121802] env[70013]: DEBUG nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1436.148347] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1436.148582] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1436.148740] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1436.148926] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1436.149086] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1436.149240] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1436.149447] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1436.149605] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1436.149769] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1436.149935] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1436.150121] env[70013]: DEBUG nova.virt.hardware [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1436.150968] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fd4312-5462-4127-b432-5562188bfe30 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.159178] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d622d0f-86fa-49ea-9a29-9342e0431f70 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.268901] env[70013]: DEBUG nova.compute.manager [req-544da84e-90e3-4184-93db-ee53a4e6c83c req-35eab36a-c84e-4390-9796-655de2ae1eae service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Received event network-vif-plugged-38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1436.269136] env[70013]: DEBUG oslo_concurrency.lockutils [req-544da84e-90e3-4184-93db-ee53a4e6c83c req-35eab36a-c84e-4390-9796-655de2ae1eae service nova] Acquiring lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1436.269350] env[70013]: DEBUG oslo_concurrency.lockutils [req-544da84e-90e3-4184-93db-ee53a4e6c83c req-35eab36a-c84e-4390-9796-655de2ae1eae service nova] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1436.269524] env[70013]: DEBUG oslo_concurrency.lockutils [req-544da84e-90e3-4184-93db-ee53a4e6c83c req-35eab36a-c84e-4390-9796-655de2ae1eae service nova] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1436.269689] env[70013]: DEBUG nova.compute.manager [req-544da84e-90e3-4184-93db-ee53a4e6c83c req-35eab36a-c84e-4390-9796-655de2ae1eae service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] No waiting events found dispatching network-vif-plugged-38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1436.269877] env[70013]: WARNING nova.compute.manager [req-544da84e-90e3-4184-93db-ee53a4e6c83c req-35eab36a-c84e-4390-9796-655de2ae1eae service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Received unexpected event network-vif-plugged-38eb2cd8-95d0-4914-b70f-32f33e352bb7 for instance with vm_state building and task_state spawning. [ 1436.353968] env[70013]: DEBUG nova.network.neutron [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Successfully updated port: 38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1436.857452] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1436.857701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1436.857997] env[70013]: DEBUG nova.network.neutron [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1437.393176] env[70013]: DEBUG nova.network.neutron [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1437.515243] env[70013]: DEBUG nova.network.neutron [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updating instance_info_cache with network_info: [{"id": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "address": "fa:16:3e:ae:c2:af", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38eb2cd8-95", "ovs_interfaceid": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1438.018597] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1438.018997] env[70013]: DEBUG nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Instance network_info: |[{"id": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "address": "fa:16:3e:ae:c2:af", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38eb2cd8-95", "ovs_interfaceid": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1438.019514] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:c2:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38eb2cd8-95d0-4914-b70f-32f33e352bb7', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1438.027542] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1438.027826] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1438.028141] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4291dc5-65d7-40ec-b68a-af14f3820533 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.049122] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1438.049122] env[70013]: value = "task-4231695" [ 1438.049122] env[70013]: _type = "Task" [ 1438.049122] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1438.057187] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231695, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1438.296094] env[70013]: DEBUG nova.compute.manager [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Received event network-changed-38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1438.296211] env[70013]: DEBUG nova.compute.manager [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Refreshing instance network info cache due to event network-changed-38eb2cd8-95d0-4914-b70f-32f33e352bb7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1438.296428] env[70013]: DEBUG oslo_concurrency.lockutils [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] Acquiring lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1438.296618] env[70013]: DEBUG oslo_concurrency.lockutils [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] Acquired lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1438.296740] env[70013]: DEBUG nova.network.neutron [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Refreshing network info cache for port 38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1438.560394] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231695, 'name': CreateVM_Task, 'duration_secs': 0.330244} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1438.560782] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1438.561282] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1438.561467] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1438.561787] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1438.562057] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56ee6770-b205-4c46-8c8e-20b55c178b26 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.567488] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1438.567488] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cb4959-5e6e-cd3a-2554-a63f2a56faad" [ 1438.567488] env[70013]: _type = "Task" [ 1438.567488] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1438.575665] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cb4959-5e6e-cd3a-2554-a63f2a56faad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1438.991681] env[70013]: DEBUG nova.network.neutron [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updated VIF entry in instance network info cache for port 38eb2cd8-95d0-4914-b70f-32f33e352bb7. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1438.992076] env[70013]: DEBUG nova.network.neutron [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updating instance_info_cache with network_info: [{"id": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "address": "fa:16:3e:ae:c2:af", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38eb2cd8-95", "ovs_interfaceid": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1439.077853] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52cb4959-5e6e-cd3a-2554-a63f2a56faad, 'name': SearchDatastore_Task, 'duration_secs': 0.009653} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1439.078172] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1439.078412] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1439.078649] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1439.078797] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1439.079011] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1439.079240] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc1a0b1f-598d-40ca-a7f8-de80f3b30660 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.088049] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1439.088248] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1439.088953] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-401ae5ab-1bb2-4703-a592-4050eeab5a94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.094476] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1439.094476] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52415c38-fb31-5ce1-8764-42e9fd0cd2e3" [ 1439.094476] env[70013]: _type = "Task" [ 1439.094476] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1439.102039] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52415c38-fb31-5ce1-8764-42e9fd0cd2e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1439.494969] env[70013]: DEBUG oslo_concurrency.lockutils [req-01ac3a4f-5393-404f-9397-0300dee9c536 req-dc782000-9595-401f-b871-0939f2fccc3e service nova] Releasing lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1439.605740] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52415c38-fb31-5ce1-8764-42e9fd0cd2e3, 'name': SearchDatastore_Task, 'duration_secs': 0.009052} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1439.606208] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b9303d1-bc9c-48bd-9228-24adbfda15a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.611755] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1439.611755] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6df00-96bb-2fca-2615-b0a5b32dfc27" [ 1439.611755] env[70013]: _type = "Task" [ 1439.611755] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1439.620957] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6df00-96bb-2fca-2615-b0a5b32dfc27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1440.122439] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f6df00-96bb-2fca-2615-b0a5b32dfc27, 'name': SearchDatastore_Task, 'duration_secs': 0.009815} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1440.122694] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1440.122952] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1440.123259] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75133d37-49d2-4243-9ec2-bc090ed72c65 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.130267] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1440.130267] env[70013]: value = "task-4231696" [ 1440.130267] env[70013]: _type = "Task" [ 1440.130267] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1440.138668] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1440.639963] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231696, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485954} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1440.640368] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1440.640443] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1440.640694] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa054b70-a3f1-468f-92ba-a8abe9294903 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.648673] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1440.648673] env[70013]: value = "task-4231697" [ 1440.648673] env[70013]: _type = "Task" [ 1440.648673] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1440.657489] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231697, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.159614] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231697, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063785} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1441.159936] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1441.160877] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa57750c-c1a4-4cf7-8d15-b196a54abba0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.184875] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1441.185142] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79eadce1-a36f-40ef-bfbd-80604872d9a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.206084] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1441.206084] env[70013]: value = "task-4231698" [ 1441.206084] env[70013]: _type = "Task" [ 1441.206084] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1441.214476] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231698, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.716441] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231698, 'name': ReconfigVM_Task, 'duration_secs': 0.265804} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1441.716884] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfigured VM instance instance-00000072 to attach disk [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1441.717450] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a8bec89-f040-4d4c-baa6-e79eaca6f40e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.724497] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1441.724497] env[70013]: value = "task-4231699" [ 1441.724497] env[70013]: _type = "Task" [ 1441.724497] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1441.733503] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231699, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1442.236126] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231699, 'name': Rename_Task, 'duration_secs': 0.204173} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.236419] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1442.236704] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d2a69b9-3e79-4c35-a12e-64369c1d1460 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.244532] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1442.244532] env[70013]: value = "task-4231700" [ 1442.244532] env[70013]: _type = "Task" [ 1442.244532] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.253140] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231700, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1442.755526] env[70013]: DEBUG oslo_vmware.api [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231700, 'name': PowerOnVM_Task, 'duration_secs': 0.45055} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.755935] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1442.756024] env[70013]: INFO nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Took 6.63 seconds to spawn the instance on the hypervisor. [ 1442.756201] env[70013]: DEBUG nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1442.756982] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d12ff8-de6c-4558-8ed3-7519679cbd2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.274068] env[70013]: INFO nova.compute.manager [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Took 11.41 seconds to build instance. [ 1443.776812] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dfac7604-e97f-4111-9985-0627726cbbdb tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.920s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1443.880906] env[70013]: DEBUG nova.compute.manager [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Received event network-changed-38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1443.881127] env[70013]: DEBUG nova.compute.manager [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Refreshing instance network info cache due to event network-changed-38eb2cd8-95d0-4914-b70f-32f33e352bb7. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1443.881370] env[70013]: DEBUG oslo_concurrency.lockutils [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] Acquiring lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1443.881514] env[70013]: DEBUG oslo_concurrency.lockutils [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] Acquired lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1443.881674] env[70013]: DEBUG nova.network.neutron [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Refreshing network info cache for port 38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1444.586584] env[70013]: DEBUG nova.network.neutron [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updated VIF entry in instance network info cache for port 38eb2cd8-95d0-4914-b70f-32f33e352bb7. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1444.586944] env[70013]: DEBUG nova.network.neutron [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updating instance_info_cache with network_info: [{"id": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "address": "fa:16:3e:ae:c2:af", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38eb2cd8-95", "ovs_interfaceid": "38eb2cd8-95d0-4914-b70f-32f33e352bb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1445.089814] env[70013]: DEBUG oslo_concurrency.lockutils [req-06152853-afd3-49ac-aaa7-57244725c4d0 req-86dce6ad-c23b-4a84-92b6-0c497a6edd99 service nova] Releasing lock "refresh_cache-e537b72b-ec82-4175-8bd9-f6ff4f98c149" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1472.573777] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1472.574196] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1472.574256] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1472.574448] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1472.574625] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1472.577407] env[70013]: INFO nova.compute.manager [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Terminating instance [ 1473.081736] env[70013]: DEBUG nova.compute.manager [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1473.082069] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1473.082986] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb78529-dabe-42ba-843c-d8070c9bf233 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.091275] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1473.091523] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8e80956-3c94-451e-83f0-b77a27200f12 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.098173] env[70013]: DEBUG oslo_vmware.api [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1473.098173] env[70013]: value = "task-4231701" [ 1473.098173] env[70013]: _type = "Task" [ 1473.098173] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1473.106238] env[70013]: DEBUG oslo_vmware.api [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231701, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1473.608704] env[70013]: DEBUG oslo_vmware.api [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231701, 'name': PowerOffVM_Task, 'duration_secs': 0.231109} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1473.609138] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1473.609188] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1473.609423] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5cb71869-912c-4cf7-92ad-110d45186d0e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.681046] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1473.681291] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1473.681518] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Deleting the datastore file [datastore2] 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1473.681824] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a56b257-26c3-4018-a23c-b59710bb9ecc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.689478] env[70013]: DEBUG oslo_vmware.api [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1473.689478] env[70013]: value = "task-4231703" [ 1473.689478] env[70013]: _type = "Task" [ 1473.689478] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1473.698472] env[70013]: DEBUG oslo_vmware.api [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1474.199817] env[70013]: DEBUG oslo_vmware.api [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136053} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1474.200030] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1474.200226] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1474.200401] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1474.200576] env[70013]: INFO nova.compute.manager [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1474.200829] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1474.201039] env[70013]: DEBUG nova.compute.manager [-] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1474.201143] env[70013]: DEBUG nova.network.neutron [-] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1474.659515] env[70013]: DEBUG nova.compute.manager [req-87f1149f-323a-4ceb-8c4b-36de6dbbd4b5 req-ae243d45-eeaf-4520-9092-abfac0b5847f service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Received event network-vif-deleted-0641d8ef-98b5-48f3-a1a8-6bdafa0154d4 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1474.659754] env[70013]: INFO nova.compute.manager [req-87f1149f-323a-4ceb-8c4b-36de6dbbd4b5 req-ae243d45-eeaf-4520-9092-abfac0b5847f service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Neutron deleted interface 0641d8ef-98b5-48f3-a1a8-6bdafa0154d4; detaching it from the instance and deleting it from the info cache [ 1474.659903] env[70013]: DEBUG nova.network.neutron [req-87f1149f-323a-4ceb-8c4b-36de6dbbd4b5 req-ae243d45-eeaf-4520-9092-abfac0b5847f service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1475.143187] env[70013]: DEBUG nova.network.neutron [-] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1475.162666] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aaec9e88-cd34-4ce4-8fbc-3f52fe3cac3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1475.175221] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a6a972-2a78-482b-ac19-7096a311ef95 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1475.201165] env[70013]: DEBUG nova.compute.manager [req-87f1149f-323a-4ceb-8c4b-36de6dbbd4b5 req-ae243d45-eeaf-4520-9092-abfac0b5847f service nova] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Detach interface failed, port_id=0641d8ef-98b5-48f3-a1a8-6bdafa0154d4, reason: Instance 8b74079f-e9d3-4212-8b7c-31f59e75ca13 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1475.645925] env[70013]: INFO nova.compute.manager [-] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Took 1.44 seconds to deallocate network for instance. [ 1476.152860] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1476.153277] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1476.153407] env[70013]: DEBUG nova.objects.instance [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'resources' on Instance uuid 8b74079f-e9d3-4212-8b7c-31f59e75ca13 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1476.788093] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab7ed03-09e4-4ba9-8064-5e758ddd721f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.795892] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800e70a1-42ce-49da-92b6-a76b6cab43ed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.827135] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63611f6c-d0db-4337-bcbd-f1dac1c5550a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.834388] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84fc37e-239d-4d47-8257-8ab1e7ea3b3d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.847671] env[70013]: DEBUG nova.compute.provider_tree [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1477.350760] env[70013]: DEBUG nova.scheduler.client.report [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1477.856262] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.703s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1477.878993] env[70013]: INFO nova.scheduler.client.report [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Deleted allocations for instance 8b74079f-e9d3-4212-8b7c-31f59e75ca13 [ 1478.387225] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2670d8c5-e3ee-4f7c-944a-2a6c1b094ca5 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "8b74079f-e9d3-4212-8b7c-31f59e75ca13" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.813s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1479.969976] env[70013]: DEBUG oslo_concurrency.lockutils [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1479.970254] env[70013]: DEBUG oslo_concurrency.lockutils [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1480.473902] env[70013]: DEBUG nova.compute.utils [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1480.504275] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1480.504498] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1480.977565] env[70013]: DEBUG oslo_concurrency.lockutils [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1481.007307] env[70013]: DEBUG nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1481.530063] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1481.530350] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1481.532450] env[70013]: INFO nova.compute.claims [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1482.040671] env[70013]: DEBUG oslo_concurrency.lockutils [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1482.040984] env[70013]: DEBUG oslo_concurrency.lockutils [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1482.041162] env[70013]: INFO nova.compute.manager [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Attaching volume 3df7ca47-2cfe-42d1-822d-321ceaaf2d54 to /dev/sdb [ 1482.071921] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75b011d-dd42-4c18-bc81-b8ea20883174 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.079460] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739cd7d3-70c6-49e7-a7b1-33fc6e7e6613 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.096955] env[70013]: DEBUG nova.virt.block_device [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updating existing volume attachment record: 1c458353-96ad-47b6-9f82-ce5a56dad619 {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1482.339693] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1482.684135] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d1b56d-c8f5-48dd-b5a7-9fee76f9269e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.691796] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07351df8-ffd1-413d-a355-35d4a907527d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.722527] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c6bd3c-3f32-44e7-ae11-066e0eaab442 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.730125] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3082bb0d-7e4c-471a-b066-3308eb88f0d1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.743153] env[70013]: DEBUG nova.compute.provider_tree [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1482.842552] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1483.246782] env[70013]: DEBUG nova.scheduler.client.report [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1483.751735] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.221s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1483.752332] env[70013]: DEBUG nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1483.755115] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.913s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1483.755309] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1483.755463] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1483.756550] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5bcb8c6-41f1-4788-a284-5a84e9ca8aa9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.765143] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86b7620-7930-4af4-8e09-0fe412ca793f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.780536] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c432d9-5732-4198-bdfd-54cdf311de4b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.788356] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adee03f-8634-447f-8d0c-f5e6b4637d82 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.818246] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179994MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1483.818452] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1483.818640] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1484.258067] env[70013]: DEBUG nova.compute.utils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1484.259606] env[70013]: DEBUG nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1484.259809] env[70013]: DEBUG nova.network.neutron [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1484.310958] env[70013]: DEBUG nova.policy [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a78bb398602948da9be11bd7ba85146d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0670a7ef480b4d39b532968ac97bbe33', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1484.619408] env[70013]: DEBUG nova.network.neutron [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Successfully created port: 085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1484.763934] env[70013]: DEBUG nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1484.850895] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851067] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851199] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851322] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851440] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851554] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851667] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851781] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.851894] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8d90b797-118d-45cb-9422-7137b7fadbe2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.852015] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance e537b72b-ec82-4175-8bd9-f6ff4f98c149 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.852144] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 89a7c112-4f31-40ff-bb7b-f6f6415289ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1484.852344] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1484.852485] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1484.989903] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc96221c-aa57-4639-801d-27d12f8ad4be {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.998017] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d3f6cd-7af3-483e-bbff-7f6073ef8191 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.028267] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ab3421-23df-46fe-9c41-79c4c6b03738 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.036206] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228fc57f-e4eb-4642-9a69-31337c433a38 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.049767] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1485.554166] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1485.775891] env[70013]: DEBUG nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1485.802051] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1485.802051] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1485.802051] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1485.802312] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1485.802350] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1485.802500] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1485.802711] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1485.802869] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1485.803076] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1485.803255] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1485.803433] env[70013]: DEBUG nova.virt.hardware [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1485.804399] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387356cf-eb67-4979-86c8-3326c21ddb1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.812948] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915f8210-f064-4d89-ab34-6abb84a1e8fb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.976083] env[70013]: DEBUG nova.compute.manager [req-7dafc216-c692-476f-99f1-ccf78f1420a9 req-dd7ebddb-86af-453d-bac3-74c2c9066b38 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Received event network-vif-plugged-085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1485.976369] env[70013]: DEBUG oslo_concurrency.lockutils [req-7dafc216-c692-476f-99f1-ccf78f1420a9 req-dd7ebddb-86af-453d-bac3-74c2c9066b38 service nova] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1485.976584] env[70013]: DEBUG oslo_concurrency.lockutils [req-7dafc216-c692-476f-99f1-ccf78f1420a9 req-dd7ebddb-86af-453d-bac3-74c2c9066b38 service nova] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1485.976756] env[70013]: DEBUG oslo_concurrency.lockutils [req-7dafc216-c692-476f-99f1-ccf78f1420a9 req-dd7ebddb-86af-453d-bac3-74c2c9066b38 service nova] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1485.976925] env[70013]: DEBUG nova.compute.manager [req-7dafc216-c692-476f-99f1-ccf78f1420a9 req-dd7ebddb-86af-453d-bac3-74c2c9066b38 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] No waiting events found dispatching network-vif-plugged-085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1485.977105] env[70013]: WARNING nova.compute.manager [req-7dafc216-c692-476f-99f1-ccf78f1420a9 req-dd7ebddb-86af-453d-bac3-74c2c9066b38 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Received unexpected event network-vif-plugged-085bf42e-3339-4556-90f6-03e04a11cfb5 for instance with vm_state building and task_state spawning. [ 1486.058613] env[70013]: DEBUG nova.network.neutron [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Successfully updated port: 085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1486.060329] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1486.060384] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.242s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1486.563610] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1486.563951] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1486.564153] env[70013]: DEBUG nova.network.neutron [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1486.640058] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1486.640351] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1486.641257] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250413eb-c78c-4df0-abc6-9fd11a1dc703 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.658056] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ab03fa-f425-4cd0-8ef5-881aa4c0fdbe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.683287] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54/volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1486.683634] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84c40ebf-ea6a-4968-85fa-546d55edafae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.702399] env[70013]: DEBUG oslo_vmware.api [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1486.702399] env[70013]: value = "task-4231708" [ 1486.702399] env[70013]: _type = "Task" [ 1486.702399] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1486.711148] env[70013]: DEBUG oslo_vmware.api [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231708, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1487.060867] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1487.061084] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1487.061310] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1487.061529] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1487.061704] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1487.095921] env[70013]: DEBUG nova.network.neutron [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1487.213851] env[70013]: DEBUG oslo_vmware.api [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231708, 'name': ReconfigVM_Task, 'duration_secs': 0.332578} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1487.214151] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfigured VM instance instance-00000072 to attach disk [datastore2] volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54/volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1487.218873] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc1519d5-b63a-410b-a81c-193650f22e20 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.230998] env[70013]: DEBUG nova.network.neutron [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updating instance_info_cache with network_info: [{"id": "085bf42e-3339-4556-90f6-03e04a11cfb5", "address": "fa:16:3e:f6:44:1b", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap085bf42e-33", "ovs_interfaceid": "085bf42e-3339-4556-90f6-03e04a11cfb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1487.234306] env[70013]: DEBUG oslo_vmware.api [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1487.234306] env[70013]: value = "task-4231709" [ 1487.234306] env[70013]: _type = "Task" [ 1487.234306] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1487.242609] env[70013]: DEBUG oslo_vmware.api [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231709, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1487.334964] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1487.339799] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1487.733800] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1487.734220] env[70013]: DEBUG nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Instance network_info: |[{"id": "085bf42e-3339-4556-90f6-03e04a11cfb5", "address": "fa:16:3e:f6:44:1b", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap085bf42e-33", "ovs_interfaceid": "085bf42e-3339-4556-90f6-03e04a11cfb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1487.734729] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:44:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '085bf42e-3339-4556-90f6-03e04a11cfb5', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1487.742145] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1487.742436] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1487.745881] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81df49ef-0962-48b8-a29b-73a76262eb5d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.767638] env[70013]: DEBUG oslo_vmware.api [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231709, 'name': ReconfigVM_Task, 'duration_secs': 0.152711} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1487.768956] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1487.770397] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1487.770397] env[70013]: value = "task-4231710" [ 1487.770397] env[70013]: _type = "Task" [ 1487.770397] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1487.779060] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231710, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1488.003636] env[70013]: DEBUG nova.compute.manager [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Received event network-changed-085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1488.003819] env[70013]: DEBUG nova.compute.manager [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Refreshing instance network info cache due to event network-changed-085bf42e-3339-4556-90f6-03e04a11cfb5. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1488.004105] env[70013]: DEBUG oslo_concurrency.lockutils [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] Acquiring lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1488.004282] env[70013]: DEBUG oslo_concurrency.lockutils [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] Acquired lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1488.004456] env[70013]: DEBUG nova.network.neutron [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Refreshing network info cache for port 085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1488.283486] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231710, 'name': CreateVM_Task, 'duration_secs': 0.313728} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1488.283663] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1488.284440] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1488.284629] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1488.284953] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1488.285241] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3f066b8-78b6-4357-a55b-69f2a784bbc9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1488.290774] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1488.290774] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52582890-2288-39ea-a077-0024252e30cd" [ 1488.290774] env[70013]: _type = "Task" [ 1488.290774] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1488.299247] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52582890-2288-39ea-a077-0024252e30cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1488.333657] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1488.691507] env[70013]: DEBUG nova.network.neutron [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updated VIF entry in instance network info cache for port 085bf42e-3339-4556-90f6-03e04a11cfb5. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1488.691974] env[70013]: DEBUG nova.network.neutron [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updating instance_info_cache with network_info: [{"id": "085bf42e-3339-4556-90f6-03e04a11cfb5", "address": "fa:16:3e:f6:44:1b", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap085bf42e-33", "ovs_interfaceid": "085bf42e-3339-4556-90f6-03e04a11cfb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1488.801919] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52582890-2288-39ea-a077-0024252e30cd, 'name': SearchDatastore_Task, 'duration_secs': 0.009975} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1488.802321] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1488.802487] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1488.802735] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1488.802864] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1488.803047] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1488.803317] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b424ff5-8fd2-44c5-b963-6fbf093db27d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1488.807468] env[70013]: DEBUG nova.objects.instance [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'flavor' on Instance uuid e537b72b-ec82-4175-8bd9-f6ff4f98c149 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1488.812662] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1488.812848] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1488.813591] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c0bb254-4108-46ed-be90-5df4d95c1045 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1488.819299] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1488.819299] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52047b2f-5730-63ce-fdaa-d6ac493faca6" [ 1488.819299] env[70013]: _type = "Task" [ 1488.819299] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1488.830369] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52047b2f-5730-63ce-fdaa-d6ac493faca6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1489.195076] env[70013]: DEBUG oslo_concurrency.lockutils [req-3a852441-994c-42db-8aac-85e5954ad38f req-d8807349-8a4e-4bb9-83f8-edeafee69839 service nova] Releasing lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1489.313212] env[70013]: DEBUG oslo_concurrency.lockutils [None req-863da51c-e1b5-4286-9616-d71716eff8f9 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.272s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1489.331034] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52047b2f-5730-63ce-fdaa-d6ac493faca6, 'name': SearchDatastore_Task, 'duration_secs': 0.01033} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1489.331845] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14ba12d4-dda0-40ea-ad6d-e986882210bf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1489.337219] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1489.337219] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bcaa0c-ceed-9786-c3e7-14e6c8e6cba2" [ 1489.337219] env[70013]: _type = "Task" [ 1489.337219] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1489.345264] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bcaa0c-ceed-9786-c3e7-14e6c8e6cba2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1489.633906] env[70013]: INFO nova.compute.manager [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Rebuilding instance [ 1489.679924] env[70013]: DEBUG nova.compute.manager [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1489.680811] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855dda9a-31e7-4dad-a34f-25c3bd307df6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1489.847783] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52bcaa0c-ceed-9786-c3e7-14e6c8e6cba2, 'name': SearchDatastore_Task, 'duration_secs': 0.009426} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1489.848457] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1489.848693] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 89a7c112-4f31-40ff-bb7b-f6f6415289ae/89a7c112-4f31-40ff-bb7b-f6f6415289ae.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1489.848957] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a3368a6-7102-4756-bbac-555daf48e820 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1489.856513] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1489.856513] env[70013]: value = "task-4231711" [ 1489.856513] env[70013]: _type = "Task" [ 1489.856513] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1489.864202] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.366564] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231711, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436897} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1490.366831] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 89a7c112-4f31-40ff-bb7b-f6f6415289ae/89a7c112-4f31-40ff-bb7b-f6f6415289ae.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1490.367071] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1490.367361] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fa63c3ef-0cbb-4b58-93a3-d9475557758e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.374460] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1490.374460] env[70013]: value = "task-4231712" [ 1490.374460] env[70013]: _type = "Task" [ 1490.374460] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.384023] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231712, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.695747] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1490.696144] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3572e37-80a5-4eab-ad85-0215b8ee61d8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.703577] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1490.703577] env[70013]: value = "task-4231713" [ 1490.703577] env[70013]: _type = "Task" [ 1490.703577] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.712083] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.884125] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231712, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06163} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1490.884501] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1490.885191] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8385c8fd-5c6c-49f5-acc9-9e4d81529b37 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.907706] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 89a7c112-4f31-40ff-bb7b-f6f6415289ae/89a7c112-4f31-40ff-bb7b-f6f6415289ae.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1490.907903] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c595c229-359b-403d-8364-cc176e618114 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.927682] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1490.927682] env[70013]: value = "task-4231714" [ 1490.927682] env[70013]: _type = "Task" [ 1490.927682] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.935623] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231714, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.212970] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.438366] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231714, 'name': ReconfigVM_Task, 'duration_secs': 0.281481} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.438694] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 89a7c112-4f31-40ff-bb7b-f6f6415289ae/89a7c112-4f31-40ff-bb7b-f6f6415289ae.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1491.439392] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-259ae5d6-ba28-4f4f-bc2d-b9b353739e55 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.445549] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1491.445549] env[70013]: value = "task-4231715" [ 1491.445549] env[70013]: _type = "Task" [ 1491.445549] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1491.455025] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231715, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.715271] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231713, 'name': PowerOffVM_Task, 'duration_secs': 0.9385} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.715593] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1491.778743] env[70013]: INFO nova.compute.manager [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Detaching volume 3df7ca47-2cfe-42d1-822d-321ceaaf2d54 [ 1491.813383] env[70013]: INFO nova.virt.block_device [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Attempting to driver detach volume 3df7ca47-2cfe-42d1-822d-321ceaaf2d54 from mountpoint /dev/sdb [ 1491.813629] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1491.813815] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1491.814828] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8cc8aa-066d-4df9-9740-1e642726e0dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.837397] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988a9ddd-31c2-402e-a7dd-34f0edb2f1bc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.845364] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502be418-e4b0-4b8e-9770-190afb12b8d6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.867673] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae117bd-0cf0-4dd3-bf7c-ff2cb6fb67f0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.883291] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] The volume has not been displaced from its original location: [datastore2] volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54/volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1491.888692] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1491.889037] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-281174d8-2c1b-4abf-9856-b67d79bff33d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.907760] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1491.907760] env[70013]: value = "task-4231716" [ 1491.907760] env[70013]: _type = "Task" [ 1491.907760] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1491.917583] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231716, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.957933] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231715, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1492.339632] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1492.418013] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231716, 'name': ReconfigVM_Task, 'duration_secs': 0.227321} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1492.418357] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1492.422990] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-776f9b71-a3f4-4d77-9e89-d725a79aa3a8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.439402] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1492.439402] env[70013]: value = "task-4231717" [ 1492.439402] env[70013]: _type = "Task" [ 1492.439402] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1492.448838] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231717, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1492.457881] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231715, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1492.949566] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231717, 'name': ReconfigVM_Task, 'duration_secs': 0.200981} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1492.954037] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1492.962592] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231715, 'name': Rename_Task, 'duration_secs': 1.15852} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1492.962886] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1492.963168] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1edec71d-d2e2-42d5-b389-6e1504b30fd3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.971255] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1492.971255] env[70013]: value = "task-4231718" [ 1492.971255] env[70013]: _type = "Task" [ 1492.971255] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1492.980912] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231718, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1493.481759] env[70013]: DEBUG oslo_vmware.api [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231718, 'name': PowerOnVM_Task, 'duration_secs': 0.460355} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1493.482049] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1493.482266] env[70013]: INFO nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1493.482451] env[70013]: DEBUG nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1493.483240] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff7e4ca-19d0-4ec2-aaef-f16c4a6d509b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.002226] env[70013]: INFO nova.compute.manager [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Took 12.49 seconds to build instance. [ 1494.003114] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1494.003535] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a368ce2-1e34-41e0-a0c5-2498566b1e0c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.011078] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1494.011078] env[70013]: value = "task-4231719" [ 1494.011078] env[70013]: _type = "Task" [ 1494.011078] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1494.019547] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1494.504673] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5f590092-0338-4318-937d-9a6750a025cb tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1494.521920] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] VM already powered off {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1494.522212] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1494.522370] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1494.523177] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58f9912-7986-4ab1-b0db-d7144e406d5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.542295] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e02ebf-b42c-4dc1-828e-cd05288eee09 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.550378] env[70013]: WARNING nova.virt.vmwareapi.driver [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1494.550778] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1494.551350] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2768674e-1919-4ee1-bcc9-22965c3996ea {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.557785] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1494.558105] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eaf3b0a3-12dc-42b3-ab8b-ff63ade873c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.617194] env[70013]: DEBUG nova.compute.manager [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Received event network-changed-085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1494.617438] env[70013]: DEBUG nova.compute.manager [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Refreshing instance network info cache due to event network-changed-085bf42e-3339-4556-90f6-03e04a11cfb5. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1494.618064] env[70013]: DEBUG oslo_concurrency.lockutils [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] Acquiring lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1494.618220] env[70013]: DEBUG oslo_concurrency.lockutils [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] Acquired lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1494.618393] env[70013]: DEBUG nova.network.neutron [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Refreshing network info cache for port 085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1494.624057] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1494.624234] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1494.624368] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleting the datastore file [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1494.624625] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c1c526d-e41d-4c59-b745-910c4f31683c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.631776] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1494.631776] env[70013]: value = "task-4231721" [ 1494.631776] env[70013]: _type = "Task" [ 1494.631776] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1494.640050] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1495.142991] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131008} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1495.142991] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1495.142991] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1495.142991] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1495.325861] env[70013]: DEBUG nova.network.neutron [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updated VIF entry in instance network info cache for port 085bf42e-3339-4556-90f6-03e04a11cfb5. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1495.326308] env[70013]: DEBUG nova.network.neutron [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updating instance_info_cache with network_info: [{"id": "085bf42e-3339-4556-90f6-03e04a11cfb5", "address": "fa:16:3e:f6:44:1b", "network": {"id": "0cf9ed9b-844e-4ac5-8804-c77a954e2955", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1110544004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0670a7ef480b4d39b532968ac97bbe33", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap085bf42e-33", "ovs_interfaceid": "085bf42e-3339-4556-90f6-03e04a11cfb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1495.647459] env[70013]: INFO nova.virt.block_device [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Booting with volume 3df7ca47-2cfe-42d1-822d-321ceaaf2d54 at /dev/sdb [ 1495.683772] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea85440e-0489-4216-8233-c5099a8dca3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.694520] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f68cc8-ebb3-4cf9-bcdb-fea10df718f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.719798] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c31a16e-454e-4e30-bee9-cf234f34647d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.728144] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7914ffbf-bea5-4466-86c1-7a32de41fc89 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.755128] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986afab6-eb48-46f6-a618-75a013ae1dd6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.761612] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988ca3bd-0a3c-45ed-947c-9a7b5b0631cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.774711] env[70013]: DEBUG nova.virt.block_device [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updating existing volume attachment record: f30261dd-99d9-4e2b-afb1-95cd853a714c {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1495.829526] env[70013]: DEBUG oslo_concurrency.lockutils [req-32c87981-bdbd-4972-903f-7869910ac922 req-019fd23f-be7e-4253-b0d5-3ddb57381b79 service nova] Releasing lock "refresh_cache-89a7c112-4f31-40ff-bb7b-f6f6415289ae" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1497.888055] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1497.888381] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1497.888381] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1497.888550] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1497.888696] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1497.888842] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1497.889056] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1497.889223] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1497.889392] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1497.889553] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1497.889722] env[70013]: DEBUG nova.virt.hardware [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1497.890605] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8b8f4e-fa80-4ad1-a2dd-076558e80116 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.899255] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8680554-8c07-4da7-ab82-20fc16f52976 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.913506] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:c2:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38eb2cd8-95d0-4914-b70f-32f33e352bb7', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1497.920889] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1497.921144] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1497.921353] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0637a22e-4048-4008-9ca5-8566bb255df9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.940943] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1497.940943] env[70013]: value = "task-4231722" [ 1497.940943] env[70013]: _type = "Task" [ 1497.940943] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1497.948480] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231722, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1498.451792] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231722, 'name': CreateVM_Task, 'duration_secs': 0.2897} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1498.452060] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1498.452835] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1498.453107] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1498.453515] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1498.453833] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df886493-52fc-45a1-a6d7-82d2b21044ef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.459239] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1498.459239] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5255e5c6-aec2-1832-50ec-f52a506a7808" [ 1498.459239] env[70013]: _type = "Task" [ 1498.459239] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1498.468063] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5255e5c6-aec2-1832-50ec-f52a506a7808, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1498.970329] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5255e5c6-aec2-1832-50ec-f52a506a7808, 'name': SearchDatastore_Task, 'duration_secs': 0.010635} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1498.970723] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1498.970865] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1498.971120] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1498.971271] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1498.971451] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1498.971723] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b62dbdef-a0dd-4814-9ffa-cf9c583068ee {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.980455] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1498.980628] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1498.981415] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f08a695-ab72-493d-8dac-5dd8a310d9d4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.987150] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1498.987150] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5207f695-4a80-7645-0489-370e3107663a" [ 1498.987150] env[70013]: _type = "Task" [ 1498.987150] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1498.994815] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5207f695-4a80-7645-0489-370e3107663a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1499.499622] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5207f695-4a80-7645-0489-370e3107663a, 'name': SearchDatastore_Task, 'duration_secs': 0.008569} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1499.500353] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-997c1b2a-c073-4e66-aa78-5d4b3b87d5aa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.505999] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1499.505999] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e3bf16-45df-f596-e749-c522f3ee12f5" [ 1499.505999] env[70013]: _type = "Task" [ 1499.505999] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1499.514052] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e3bf16-45df-f596-e749-c522f3ee12f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1500.016623] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e3bf16-45df-f596-e749-c522f3ee12f5, 'name': SearchDatastore_Task, 'duration_secs': 0.009766} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.017019] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1500.017252] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1500.017614] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-387e1540-c2d6-4ed4-a3cc-ef8c67368422 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.024689] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1500.024689] env[70013]: value = "task-4231723" [ 1500.024689] env[70013]: _type = "Task" [ 1500.024689] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1500.033047] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231723, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1500.534946] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231723, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455939} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.535227] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1500.535371] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1500.535639] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c7d6843-4db0-4bfc-afba-b267938bd9b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.543000] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1500.543000] env[70013]: value = "task-4231724" [ 1500.543000] env[70013]: _type = "Task" [ 1500.543000] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1500.552378] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231724, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1501.053487] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063267} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1501.053881] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1501.054550] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe34767-4d9e-4e65-a3ee-9d40270cb812 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.077878] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1501.078196] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aebfbd7d-bce3-4bec-bea6-5231d6664117 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.099036] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1501.099036] env[70013]: value = "task-4231725" [ 1501.099036] env[70013]: _type = "Task" [ 1501.099036] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1501.107865] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231725, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1501.609556] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231725, 'name': ReconfigVM_Task, 'duration_secs': 0.336977} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1501.609835] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfigured VM instance instance-00000072 to attach disk [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149/e537b72b-ec82-4175-8bd9-f6ff4f98c149.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1501.610949] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encrypted': False, 'encryption_secret_uuid': None, 'device_type': 'disk', 'boot_index': 0, 'encryption_options': None, 'disk_bus': None, 'size': 0, 'encryption_format': None, 'device_name': '/dev/sda', 'image_id': '0ddeef65-a14c-4065-97bb-58b607968d19'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'f30261dd-99d9-4e2b-afb1-95cd853a714c', 'guest_format': None, 'device_type': None, 'mount_device': '/dev/sdb', 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'}, 'disk_bus': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=70013) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1501.611167] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1501.611392] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1501.612196] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ccadb4-70d4-4791-8268-03c23eb07f69 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.634027] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78d909a-b148-48ec-b019-cab4671970e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.658849] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54/volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1501.659205] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30f7ef68-8f99-412d-adef-62a0126c8af2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.677514] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1501.677514] env[70013]: value = "task-4231726" [ 1501.677514] env[70013]: _type = "Task" [ 1501.677514] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1501.687249] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231726, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1502.188277] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231726, 'name': ReconfigVM_Task, 'duration_secs': 0.356963} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1502.188675] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfigured VM instance instance-00000072 to attach disk [datastore2] volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54/volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1502.193110] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5709154b-a101-4001-9fda-57a0a138f37c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1502.208732] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1502.208732] env[70013]: value = "task-4231727" [ 1502.208732] env[70013]: _type = "Task" [ 1502.208732] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1502.217971] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231727, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1502.718954] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231727, 'name': ReconfigVM_Task, 'duration_secs': 0.15601} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1502.719231] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1502.719825] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0c726ea-23d2-457d-b60c-50c6758ea9fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1502.726553] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1502.726553] env[70013]: value = "task-4231728" [ 1502.726553] env[70013]: _type = "Task" [ 1502.726553] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1502.734762] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231728, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1503.237657] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231728, 'name': Rename_Task, 'duration_secs': 0.151447} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1503.238170] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1503.238278] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3776a809-5ac3-4111-a039-fad7250cf64e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.245909] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1503.245909] env[70013]: value = "task-4231729" [ 1503.245909] env[70013]: _type = "Task" [ 1503.245909] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1503.254492] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1503.758578] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231729, 'name': PowerOnVM_Task} progress is 88%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1504.257176] env[70013]: DEBUG oslo_vmware.api [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231729, 'name': PowerOnVM_Task, 'duration_secs': 0.614006} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1504.257556] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1504.257685] env[70013]: DEBUG nova.compute.manager [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1504.258453] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e6888f-39ee-4726-901b-604450e97234 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1504.777743] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1504.778075] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1504.778271] env[70013]: DEBUG nova.objects.instance [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1505.788663] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f59552a3-d61b-492b-a91a-00f32585cf68 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1531.158427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1531.158744] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1531.663636] env[70013]: DEBUG nova.compute.utils [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1532.167168] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1533.221643] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1533.222109] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1533.222205] env[70013]: INFO nova.compute.manager [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Attaching volume 722c2558-1002-458f-99fa-59dddf9ad7a6 to /dev/sdb [ 1533.254162] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9bb4c9-6df3-4f36-aad4-ccacf8311c6d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.261988] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbd4f68-f523-4aed-93ff-afb925c9c6a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.275193] env[70013]: DEBUG nova.virt.block_device [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updating existing volume attachment record: 7fe93f54-2aef-42a3-9724-1df14b8f29e1 {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1537.819056] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1537.819315] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837256', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'name': 'volume-722c2558-1002-458f-99fa-59dddf9ad7a6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'serial': '722c2558-1002-458f-99fa-59dddf9ad7a6'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1537.820193] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f406e6b-dcdc-43f7-b761-d13d9ebdd574 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1537.836900] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4ef926-a304-4993-8420-0f5d0811589b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1537.861583] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] volume-722c2558-1002-458f-99fa-59dddf9ad7a6/volume-722c2558-1002-458f-99fa-59dddf9ad7a6.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1537.861866] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea9d1f25-fb7b-46a2-941d-c50ab90f2815 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1537.880357] env[70013]: DEBUG oslo_vmware.api [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1537.880357] env[70013]: value = "task-4231732" [ 1537.880357] env[70013]: _type = "Task" [ 1537.880357] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1538.392617] env[70013]: DEBUG oslo_vmware.api [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231732, 'name': ReconfigVM_Task, 'duration_secs': 0.41123} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1538.392892] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfigured VM instance instance-00000073 to attach disk [datastore1] volume-722c2558-1002-458f-99fa-59dddf9ad7a6/volume-722c2558-1002-458f-99fa-59dddf9ad7a6.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1538.397427] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81aa3e76-5f43-4493-8720-89cdeeddb5c6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.412026] env[70013]: DEBUG oslo_vmware.api [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1538.412026] env[70013]: value = "task-4231733" [ 1538.412026] env[70013]: _type = "Task" [ 1538.412026] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1538.419849] env[70013]: DEBUG oslo_vmware.api [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231733, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1538.922288] env[70013]: DEBUG oslo_vmware.api [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231733, 'name': ReconfigVM_Task, 'duration_secs': 0.210366} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1538.922660] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837256', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'name': 'volume-722c2558-1002-458f-99fa-59dddf9ad7a6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'serial': '722c2558-1002-458f-99fa-59dddf9ad7a6'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1539.958349] env[70013]: DEBUG nova.objects.instance [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 89a7c112-4f31-40ff-bb7b-f6f6415289ae {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1540.464615] env[70013]: DEBUG oslo_concurrency.lockutils [None req-2b51871f-2db0-4efd-87cc-8b10ffb2aa5e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.243s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1540.912184] env[70013]: DEBUG oslo_concurrency.lockutils [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1540.912445] env[70013]: DEBUG oslo_concurrency.lockutils [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1541.347355] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1541.347658] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1541.415719] env[70013]: INFO nova.compute.manager [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Detaching volume 3df7ca47-2cfe-42d1-822d-321ceaaf2d54 [ 1541.446625] env[70013]: INFO nova.virt.block_device [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Attempting to driver detach volume 3df7ca47-2cfe-42d1-822d-321ceaaf2d54 from mountpoint /dev/sdb [ 1541.446905] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1541.447120] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1541.448065] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255c5ea2-4f89-4739-8114-006564909b2e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.470422] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c332bc0f-78d7-421e-9efb-66c7402c8004 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.479039] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99406a7d-97c1-4ca4-bb8e-d8c078adb898 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.500364] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e658de8a-770c-4aae-9298-c9f357131d9d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.516308] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] The volume has not been displaced from its original location: [datastore2] volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54/volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1541.521586] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1541.521912] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3b96ad2-46a8-4eb9-b4f7-c846e2299f0c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.540084] env[70013]: DEBUG oslo_vmware.api [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1541.540084] env[70013]: value = "task-4231734" [ 1541.540084] env[70013]: _type = "Task" [ 1541.540084] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1541.548389] env[70013]: DEBUG oslo_vmware.api [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231734, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1541.851211] env[70013]: DEBUG nova.compute.utils [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1542.050211] env[70013]: DEBUG oslo_vmware.api [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231734, 'name': ReconfigVM_Task, 'duration_secs': 0.250554} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1542.050524] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1542.055899] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-819c8959-f05c-4eab-a48c-5af14c108ec2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.076524] env[70013]: DEBUG oslo_vmware.api [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1542.076524] env[70013]: value = "task-4231735" [ 1542.076524] env[70013]: _type = "Task" [ 1542.076524] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1542.085056] env[70013]: DEBUG oslo_vmware.api [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231735, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1542.339586] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1542.353688] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1542.587486] env[70013]: DEBUG oslo_vmware.api [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231735, 'name': ReconfigVM_Task, 'duration_secs': 0.140972} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1542.587803] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837253', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'name': 'volume-3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e537b72b-ec82-4175-8bd9-f6ff4f98c149', 'attached_at': '', 'detached_at': '', 'volume_id': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54', 'serial': '3df7ca47-2cfe-42d1-822d-321ceaaf2d54'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1542.843082] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1542.843386] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1542.843494] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1542.843617] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1542.844534] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9b636f-087a-4478-9b5d-1cce37ffec64 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.853362] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4372a15d-1170-4162-8669-1da488571c59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.870529] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e54e5e-f61b-4679-870a-e4241eef2d87 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.877969] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fee7b7-742d-4b0c-86ba-de86aa77aacb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.909172] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179930MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1542.909415] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1542.909646] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1543.128739] env[70013]: DEBUG nova.objects.instance [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'flavor' on Instance uuid e537b72b-ec82-4175-8bd9-f6ff4f98c149 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1543.418142] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1543.418504] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1543.418552] env[70013]: INFO nova.compute.manager [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Attaching volume 540ae424-2740-4d97-a3a7-6267830b407c to /dev/sdc [ 1543.449150] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e4d98b-839a-4bf5-832c-8e565b3f3154 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.455908] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8695cc22-6316-42fc-8060-bcbfe5487fe7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.469201] env[70013]: DEBUG nova.virt.block_device [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updating existing volume attachment record: 5212d752-c6d5-4c61-842d-680a16eea3ac {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1543.944052] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944052] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944211] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944211] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944367] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944549] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944765] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944834] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.944915] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8d90b797-118d-45cb-9422-7137b7fadbe2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.945055] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance e537b72b-ec82-4175-8bd9-f6ff4f98c149 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.945220] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 89a7c112-4f31-40ff-bb7b-f6f6415289ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1543.945453] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1543.945756] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1544.089382] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba3ebbb-a458-46d5-86d3-d9373491f699 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.096862] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85dbea2e-3fb2-4097-a41c-1e6da5e34c80 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.126318] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06816b01-4b59-46fe-a840-523b210371c8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.133575] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151a2c8d-4ac6-4783-b719-e5eec1fbfc8b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.137497] env[70013]: DEBUG oslo_concurrency.lockutils [None req-291929d9-0062-4a47-a0be-9d9f445873a4 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.225s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1544.148204] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1544.651323] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1545.156457] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1545.156693] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.247s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1545.160382] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1545.160612] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1545.160813] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1545.161010] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1545.161193] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1545.163429] env[70013]: INFO nova.compute.manager [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Terminating instance [ 1545.667612] env[70013]: DEBUG nova.compute.manager [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1545.667612] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1545.668447] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b541c7-cf32-4cf5-98f1-ffcb9f871d89 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1545.675710] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1545.675977] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3190a71b-84a3-4f56-a86a-293ae3b1a1b1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1545.681191] env[70013]: DEBUG oslo_vmware.api [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1545.681191] env[70013]: value = "task-4231738" [ 1545.681191] env[70013]: _type = "Task" [ 1545.681191] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1545.691110] env[70013]: DEBUG oslo_vmware.api [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231738, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1546.191567] env[70013]: DEBUG oslo_vmware.api [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231738, 'name': PowerOffVM_Task, 'duration_secs': 0.213643} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1546.191766] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1546.191934] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1546.192208] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39e382ff-9bc6-4077-8810-16772ca9f045 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.258799] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1546.259030] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1546.259256] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleting the datastore file [datastore2] e537b72b-ec82-4175-8bd9-f6ff4f98c149 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1546.259529] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a176e6ee-8e5c-415b-8155-dc4c6e4e500d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.266135] env[70013]: DEBUG oslo_vmware.api [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1546.266135] env[70013]: value = "task-4231740" [ 1546.266135] env[70013]: _type = "Task" [ 1546.266135] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1546.274151] env[70013]: DEBUG oslo_vmware.api [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1546.776345] env[70013]: DEBUG oslo_vmware.api [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142578} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1546.776763] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1546.777012] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1546.777385] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1546.777689] env[70013]: INFO nova.compute.manager [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1546.778032] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1546.778280] env[70013]: DEBUG nova.compute.manager [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1546.778436] env[70013]: DEBUG nova.network.neutron [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1547.157461] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1547.157764] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1547.255059] env[70013]: DEBUG nova.compute.manager [req-885d8efa-1ec4-40c5-a0f1-a78a35e9140c req-90227193-4dce-4519-98a0-b7e8834048e6 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Received event network-vif-deleted-38eb2cd8-95d0-4914-b70f-32f33e352bb7 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1547.255191] env[70013]: INFO nova.compute.manager [req-885d8efa-1ec4-40c5-a0f1-a78a35e9140c req-90227193-4dce-4519-98a0-b7e8834048e6 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Neutron deleted interface 38eb2cd8-95d0-4914-b70f-32f33e352bb7; detaching it from the instance and deleting it from the info cache [ 1547.255337] env[70013]: DEBUG nova.network.neutron [req-885d8efa-1ec4-40c5-a0f1-a78a35e9140c req-90227193-4dce-4519-98a0-b7e8834048e6 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1547.340372] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1547.340672] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1547.340941] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1547.341245] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1547.341530] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Cleaning up deleted instances with incomplete migration {{(pid=70013) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11902}} [ 1547.738095] env[70013]: DEBUG nova.network.neutron [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1547.757495] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c48a748-0c43-4ec0-b1d6-67217d39b8a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.768666] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f750c3c-4bca-43bf-bdb5-b9b769730993 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.795351] env[70013]: DEBUG nova.compute.manager [req-885d8efa-1ec4-40c5-a0f1-a78a35e9140c req-90227193-4dce-4519-98a0-b7e8834048e6 service nova] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Detach interface failed, port_id=38eb2cd8-95d0-4914-b70f-32f33e352bb7, reason: Instance e537b72b-ec82-4175-8bd9-f6ff4f98c149 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1548.011777] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1548.012037] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837257', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'name': 'volume-540ae424-2740-4d97-a3a7-6267830b407c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'serial': '540ae424-2740-4d97-a3a7-6267830b407c'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1548.012898] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5ffb83-5b5b-405f-ad7f-b078acbf856a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1548.030487] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b98906-5ff4-4273-b4cf-251c5fde5292 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1548.062589] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] volume-540ae424-2740-4d97-a3a7-6267830b407c/volume-540ae424-2740-4d97-a3a7-6267830b407c.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1548.062942] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d265b0f-75fb-4575-8184-b13181d1d2b3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1548.084342] env[70013]: DEBUG oslo_vmware.api [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1548.084342] env[70013]: value = "task-4231741" [ 1548.084342] env[70013]: _type = "Task" [ 1548.084342] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1548.093615] env[70013]: DEBUG oslo_vmware.api [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1548.241607] env[70013]: INFO nova.compute.manager [-] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Took 1.46 seconds to deallocate network for instance. [ 1548.594308] env[70013]: DEBUG oslo_vmware.api [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231741, 'name': ReconfigVM_Task, 'duration_secs': 0.36582} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1548.594537] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfigured VM instance instance-00000073 to attach disk [datastore1] volume-540ae424-2740-4d97-a3a7-6267830b407c/volume-540ae424-2740-4d97-a3a7-6267830b407c.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1548.599276] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e71a967-9d22-4676-9049-c741c4baf7c3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1548.613693] env[70013]: DEBUG oslo_vmware.api [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1548.613693] env[70013]: value = "task-4231742" [ 1548.613693] env[70013]: _type = "Task" [ 1548.613693] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1548.621770] env[70013]: DEBUG oslo_vmware.api [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1548.748345] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1548.748605] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1548.748832] env[70013]: DEBUG nova.objects.instance [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'resources' on Instance uuid e537b72b-ec82-4175-8bd9-f6ff4f98c149 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1548.840756] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1548.841184] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1549.124759] env[70013]: DEBUG oslo_vmware.api [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231742, 'name': ReconfigVM_Task, 'duration_secs': 0.138767} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1549.125091] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837257', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'name': 'volume-540ae424-2740-4d97-a3a7-6267830b407c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'serial': '540ae424-2740-4d97-a3a7-6267830b407c'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1549.451822] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb38fe9-4e11-468d-8ef6-5f813948c746 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1549.459626] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd02951-18a2-49f0-9192-2f24e664a748 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1549.490628] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c9b036-66c3-4298-8fb7-5e035a443648 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1549.498394] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91f553a-105c-411a-8c22-b7bac9c818dd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1549.511836] env[70013]: DEBUG nova.compute.provider_tree [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1550.015122] env[70013]: DEBUG nova.scheduler.client.report [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1550.161142] env[70013]: DEBUG nova.objects.instance [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 89a7c112-4f31-40ff-bb7b-f6f6415289ae {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1550.340232] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1550.519876] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.771s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1550.541819] env[70013]: INFO nova.scheduler.client.report [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted allocations for instance e537b72b-ec82-4175-8bd9-f6ff4f98c149 [ 1550.666107] env[70013]: DEBUG oslo_concurrency.lockutils [None req-32b0ba9e-b955-4fa4-b0ac-ce5cf5c88f78 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.248s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1550.955532] env[70013]: DEBUG oslo_concurrency.lockutils [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1550.955736] env[70013]: DEBUG oslo_concurrency.lockutils [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1551.048884] env[70013]: DEBUG oslo_concurrency.lockutils [None req-f3819b64-3c0e-4133-a2fc-27c2d06f709a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "e537b72b-ec82-4175-8bd9-f6ff4f98c149" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.888s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1551.459132] env[70013]: INFO nova.compute.manager [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Detaching volume 722c2558-1002-458f-99fa-59dddf9ad7a6 [ 1551.495141] env[70013]: INFO nova.virt.block_device [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Attempting to driver detach volume 722c2558-1002-458f-99fa-59dddf9ad7a6 from mountpoint /dev/sdb [ 1551.495141] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1551.495141] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837256', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'name': 'volume-722c2558-1002-458f-99fa-59dddf9ad7a6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'serial': '722c2558-1002-458f-99fa-59dddf9ad7a6'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1551.495527] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ce67b8-fe37-419d-802f-4cfcdd1b83f6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1551.521427] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285c946c-b756-4780-9016-ce093fd926f5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1551.529153] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47aea06c-c207-499f-a5ff-7970734947c2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1551.552885] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2aae3df-1a56-47e2-968f-cd7f33f9a13b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1551.568799] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] The volume has not been displaced from its original location: [datastore1] volume-722c2558-1002-458f-99fa-59dddf9ad7a6/volume-722c2558-1002-458f-99fa-59dddf9ad7a6.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1551.574123] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfiguring VM instance instance-00000073 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1551.574762] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10268db1-e8c3-499c-a2fd-776e0cc6a572 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1551.593837] env[70013]: DEBUG oslo_vmware.api [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1551.593837] env[70013]: value = "task-4231743" [ 1551.593837] env[70013]: _type = "Task" [ 1551.593837] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1551.602534] env[70013]: DEBUG oslo_vmware.api [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231743, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1552.104255] env[70013]: DEBUG oslo_vmware.api [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231743, 'name': ReconfigVM_Task, 'duration_secs': 0.23636} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1552.104688] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfigured VM instance instance-00000073 to detach disk 2001 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1552.109315] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0bd3d33-a7d2-4b35-8c11-b1d9fc3eb66a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1552.125737] env[70013]: DEBUG oslo_vmware.api [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1552.125737] env[70013]: value = "task-4231744" [ 1552.125737] env[70013]: _type = "Task" [ 1552.125737] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1552.134544] env[70013]: DEBUG oslo_vmware.api [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231744, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1552.636134] env[70013]: DEBUG oslo_vmware.api [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231744, 'name': ReconfigVM_Task, 'duration_secs': 0.144963} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1552.636475] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837256', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'name': 'volume-722c2558-1002-458f-99fa-59dddf9ad7a6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '722c2558-1002-458f-99fa-59dddf9ad7a6', 'serial': '722c2558-1002-458f-99fa-59dddf9ad7a6'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1552.842527] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1552.854655] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1552.854982] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1553.176715] env[70013]: DEBUG nova.objects.instance [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 89a7c112-4f31-40ff-bb7b-f6f6415289ae {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1553.357767] env[70013]: DEBUG nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1553.878016] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1553.878325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1553.879801] env[70013]: INFO nova.compute.claims [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1554.185035] env[70013]: DEBUG oslo_concurrency.lockutils [None req-913879e3-8e78-4dd1-ae21-ef34dae0df1e tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.229s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1554.213680] env[70013]: DEBUG oslo_concurrency.lockutils [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1554.213990] env[70013]: DEBUG oslo_concurrency.lockutils [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1554.340472] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1554.340683] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Cleaning up deleted instances {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11864}} [ 1554.717390] env[70013]: INFO nova.compute.manager [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Detaching volume 540ae424-2740-4d97-a3a7-6267830b407c [ 1554.747040] env[70013]: INFO nova.virt.block_device [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Attempting to driver detach volume 540ae424-2740-4d97-a3a7-6267830b407c from mountpoint /dev/sdc [ 1554.747305] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1554.747515] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837257', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'name': 'volume-540ae424-2740-4d97-a3a7-6267830b407c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'serial': '540ae424-2740-4d97-a3a7-6267830b407c'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1554.748516] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7d5760-cbdb-4f6e-ac45-b3fab885ebd1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.770271] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2ff8a0-aa6a-48d9-88e9-49c77fe6fe23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.777779] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9983ba2c-7c44-40e0-aff3-00f3a0c32a73 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.798783] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880b35ad-7b31-4a48-a73a-28e0f98d7484 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.813187] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] The volume has not been displaced from its original location: [datastore1] volume-540ae424-2740-4d97-a3a7-6267830b407c/volume-540ae424-2740-4d97-a3a7-6267830b407c.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1554.818296] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfiguring VM instance instance-00000073 to detach disk 2002 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1554.818580] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98dd9267-22f4-4181-818c-28326845ab0d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.836077] env[70013]: DEBUG oslo_vmware.api [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1554.836077] env[70013]: value = "task-4231745" [ 1554.836077] env[70013]: _type = "Task" [ 1554.836077] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1554.846377] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] There are 6 instances to clean {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11873}} [ 1554.846623] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: e537b72b-ec82-4175-8bd9-f6ff4f98c149] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1554.849082] env[70013]: DEBUG oslo_vmware.api [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231745, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1555.029339] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-addfd817-c076-4c61-a5f1-1548b674d5ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.036617] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ae2311-d536-4a03-b99c-75f5e068d303 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.066819] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce664ab-b492-41ba-bf8a-e2ddcccc187e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.074348] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2075b6cb-77dc-4679-bee3-d0cb66148ffe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.089162] env[70013]: DEBUG nova.compute.provider_tree [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1555.346948] env[70013]: DEBUG oslo_vmware.api [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231745, 'name': ReconfigVM_Task, 'duration_secs': 0.221198} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1555.347270] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Reconfigured VM instance instance-00000073 to detach disk 2002 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1555.351977] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 2f7b0d3d-4e6f-458a-81af-73a525f33df3] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1555.353908] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d20c7058-0500-4450-b829-84dfb244769c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.371561] env[70013]: DEBUG oslo_vmware.api [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1555.371561] env[70013]: value = "task-4231746" [ 1555.371561] env[70013]: _type = "Task" [ 1555.371561] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1555.379995] env[70013]: DEBUG oslo_vmware.api [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231746, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1555.592888] env[70013]: DEBUG nova.scheduler.client.report [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1555.865877] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 8b74079f-e9d3-4212-8b7c-31f59e75ca13] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1555.881946] env[70013]: DEBUG oslo_vmware.api [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231746, 'name': ReconfigVM_Task, 'duration_secs': 0.138232} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1555.882327] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837257', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'name': 'volume-540ae424-2740-4d97-a3a7-6267830b407c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '89a7c112-4f31-40ff-bb7b-f6f6415289ae', 'attached_at': '', 'detached_at': '', 'volume_id': '540ae424-2740-4d97-a3a7-6267830b407c', 'serial': '540ae424-2740-4d97-a3a7-6267830b407c'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1556.098944] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.220s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1556.098944] env[70013]: DEBUG nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1556.369821] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 3d92b939-18d2-450c-b815-e0266542fbcb] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1556.422148] env[70013]: DEBUG nova.objects.instance [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'flavor' on Instance uuid 89a7c112-4f31-40ff-bb7b-f6f6415289ae {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1556.604602] env[70013]: DEBUG nova.compute.utils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1556.606100] env[70013]: DEBUG nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1556.606310] env[70013]: DEBUG nova.network.neutron [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1556.680984] env[70013]: DEBUG nova.policy [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ade87067e834404adc174783ff7b03d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66804721098f48d69e216807291268d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1556.872745] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 351f5ef4-ed63-4d5d-b0a1-c029393898ac] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1556.941758] env[70013]: DEBUG nova.network.neutron [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Successfully created port: 2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1557.109328] env[70013]: DEBUG nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1557.376027] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 96a5b996-7b8b-4a3d-a815-5c7dcf1090d6] Instance has had 0 of 5 cleanup attempts {{(pid=70013) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11877}} [ 1557.428803] env[70013]: DEBUG oslo_concurrency.lockutils [None req-13e3383d-afd3-40b9-8b8f-5a5417a3a647 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.215s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1558.118945] env[70013]: DEBUG nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1558.145701] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1558.145949] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1558.146125] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1558.146313] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1558.146462] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1558.146610] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1558.146893] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1558.147134] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1558.147356] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1558.147543] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1558.147720] env[70013]: DEBUG nova.virt.hardware [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1558.148751] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93370144-fdf5-4aa4-bc69-0ab19926859b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1558.158067] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820d1f11-ec39-46da-9bf5-b88cb071e0a2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1558.308116] env[70013]: DEBUG nova.compute.manager [req-6b08470b-0e8c-4f81-9a72-20e38e3fb1a7 req-4a20f843-b184-4b57-a9b7-764e5cb74424 service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Received event network-vif-plugged-2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1558.308425] env[70013]: DEBUG oslo_concurrency.lockutils [req-6b08470b-0e8c-4f81-9a72-20e38e3fb1a7 req-4a20f843-b184-4b57-a9b7-764e5cb74424 service nova] Acquiring lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1558.308643] env[70013]: DEBUG oslo_concurrency.lockutils [req-6b08470b-0e8c-4f81-9a72-20e38e3fb1a7 req-4a20f843-b184-4b57-a9b7-764e5cb74424 service nova] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1558.308814] env[70013]: DEBUG oslo_concurrency.lockutils [req-6b08470b-0e8c-4f81-9a72-20e38e3fb1a7 req-4a20f843-b184-4b57-a9b7-764e5cb74424 service nova] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1558.308983] env[70013]: DEBUG nova.compute.manager [req-6b08470b-0e8c-4f81-9a72-20e38e3fb1a7 req-4a20f843-b184-4b57-a9b7-764e5cb74424 service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] No waiting events found dispatching network-vif-plugged-2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1558.309317] env[70013]: WARNING nova.compute.manager [req-6b08470b-0e8c-4f81-9a72-20e38e3fb1a7 req-4a20f843-b184-4b57-a9b7-764e5cb74424 service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Received unexpected event network-vif-plugged-2b1ad1e3-7054-4541-8a7d-1987779f9a9a for instance with vm_state building and task_state spawning. [ 1558.386954] env[70013]: DEBUG nova.network.neutron [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Successfully updated port: 2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1558.599373] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1558.599648] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1558.600305] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1558.600526] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1558.600709] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1558.602931] env[70013]: INFO nova.compute.manager [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Terminating instance [ 1558.890372] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1558.890563] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1558.890748] env[70013]: DEBUG nova.network.neutron [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1559.107122] env[70013]: DEBUG nova.compute.manager [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1559.107369] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1559.108305] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab40f06-e8fd-4303-a6b9-87e858b11a1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1559.116901] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1559.117208] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cda11e3-a165-4340-8536-7c0a1e75f1d9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1559.124344] env[70013]: DEBUG oslo_vmware.api [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1559.124344] env[70013]: value = "task-4231747" [ 1559.124344] env[70013]: _type = "Task" [ 1559.124344] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1559.133208] env[70013]: DEBUG oslo_vmware.api [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231747, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1559.422891] env[70013]: DEBUG nova.network.neutron [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1559.552955] env[70013]: DEBUG nova.network.neutron [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Updating instance_info_cache with network_info: [{"id": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "address": "fa:16:3e:ae:17:55", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b1ad1e3-70", "ovs_interfaceid": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1559.634783] env[70013]: DEBUG oslo_vmware.api [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231747, 'name': PowerOffVM_Task, 'duration_secs': 0.18386} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1559.635102] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1559.635327] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1559.635631] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f969a956-40f6-4dba-a7b2-a5cf1302187e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1559.702311] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1559.702522] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1559.702708] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Deleting the datastore file [datastore2] 89a7c112-4f31-40ff-bb7b-f6f6415289ae {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1559.703010] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c6f640d-33af-4a73-bf91-aeea11c2bb87 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1559.710362] env[70013]: DEBUG oslo_vmware.api [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for the task: (returnval){ [ 1559.710362] env[70013]: value = "task-4231749" [ 1559.710362] env[70013]: _type = "Task" [ 1559.710362] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1559.718445] env[70013]: DEBUG oslo_vmware.api [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1560.056283] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1560.056597] env[70013]: DEBUG nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Instance network_info: |[{"id": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "address": "fa:16:3e:ae:17:55", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b1ad1e3-70", "ovs_interfaceid": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1560.057136] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:17:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b1ad1e3-7054-4541-8a7d-1987779f9a9a', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1560.064649] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1560.064900] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1560.065159] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4311d435-75c2-4e00-ae35-6b8b92e87cd2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1560.086488] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1560.086488] env[70013]: value = "task-4231750" [ 1560.086488] env[70013]: _type = "Task" [ 1560.086488] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1560.094529] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231750, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1560.221093] env[70013]: DEBUG oslo_vmware.api [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Task: {'id': task-4231749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141754} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1560.221394] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1560.221597] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1560.221783] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1560.221974] env[70013]: INFO nova.compute.manager [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1560.222254] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1560.222464] env[70013]: DEBUG nova.compute.manager [-] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1560.222553] env[70013]: DEBUG nova.network.neutron [-] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1560.335142] env[70013]: DEBUG nova.compute.manager [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Received event network-changed-2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1560.335350] env[70013]: DEBUG nova.compute.manager [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Refreshing instance network info cache due to event network-changed-2b1ad1e3-7054-4541-8a7d-1987779f9a9a. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1560.335577] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] Acquiring lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1560.335725] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] Acquired lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1560.335896] env[70013]: DEBUG nova.network.neutron [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Refreshing network info cache for port 2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1560.598347] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231750, 'name': CreateVM_Task, 'duration_secs': 0.402691} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1560.598785] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1560.599260] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1560.599427] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1560.599767] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1560.600256] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-142b9369-950a-4f77-bcff-23c621aa784b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1560.606219] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1560.606219] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d8e15d-27cd-f86f-5724-6279da6bd664" [ 1560.606219] env[70013]: _type = "Task" [ 1560.606219] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1560.614726] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d8e15d-27cd-f86f-5724-6279da6bd664, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1561.048900] env[70013]: DEBUG nova.network.neutron [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Updated VIF entry in instance network info cache for port 2b1ad1e3-7054-4541-8a7d-1987779f9a9a. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1561.049376] env[70013]: DEBUG nova.network.neutron [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Updating instance_info_cache with network_info: [{"id": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "address": "fa:16:3e:ae:17:55", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b1ad1e3-70", "ovs_interfaceid": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1561.117192] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d8e15d-27cd-f86f-5724-6279da6bd664, 'name': SearchDatastore_Task, 'duration_secs': 0.012949} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1561.117464] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1561.117706] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1561.117958] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1561.118125] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1561.118312] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1561.118571] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d49396a-4e96-4fa7-85ed-32c0b1fae252 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1561.127230] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1561.127416] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1561.128130] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83df4b8d-bdbf-4df9-8a8b-ea95f9644f92 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1561.133413] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1561.133413] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52316932-799c-c60a-d13d-aa44e53c060c" [ 1561.133413] env[70013]: _type = "Task" [ 1561.133413] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1561.141300] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52316932-799c-c60a-d13d-aa44e53c060c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1561.392996] env[70013]: DEBUG nova.network.neutron [-] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1561.552895] env[70013]: DEBUG oslo_concurrency.lockutils [req-f6d356ff-7f90-4446-8493-1771dfda37ad req-42bc470a-dcdb-45fe-bf1d-bd883d0a087c service nova] Releasing lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1561.644472] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52316932-799c-c60a-d13d-aa44e53c060c, 'name': SearchDatastore_Task, 'duration_secs': 0.009913} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1561.645209] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a91e4f7-f471-4eb5-b5cb-0ea10bf631f8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1561.650261] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1561.650261] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5201ad22-de4b-5310-4398-40aa5f76fad0" [ 1561.650261] env[70013]: _type = "Task" [ 1561.650261] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1561.658048] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5201ad22-de4b-5310-4398-40aa5f76fad0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1561.896465] env[70013]: INFO nova.compute.manager [-] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Took 1.67 seconds to deallocate network for instance. [ 1562.162605] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5201ad22-de4b-5310-4398-40aa5f76fad0, 'name': SearchDatastore_Task, 'duration_secs': 0.0103} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1562.162953] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1562.163291] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 6b329675-3273-4f8a-8e60-5be0d3f8a19e/6b329675-3273-4f8a-8e60-5be0d3f8a19e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1562.163574] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e734519-59d1-4fd3-a252-64c212949f75 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1562.171359] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1562.171359] env[70013]: value = "task-4231751" [ 1562.171359] env[70013]: _type = "Task" [ 1562.171359] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1562.179576] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1562.364181] env[70013]: DEBUG nova.compute.manager [req-c097f8d8-77a9-4aed-9794-ee1939833139 req-e991c0d9-d71b-4a7d-8656-afbd2d1aa0f8 service nova] [instance: 89a7c112-4f31-40ff-bb7b-f6f6415289ae] Received event network-vif-deleted-085bf42e-3339-4556-90f6-03e04a11cfb5 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1562.402809] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1562.403132] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1562.403367] env[70013]: DEBUG nova.objects.instance [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lazy-loading 'resources' on Instance uuid 89a7c112-4f31-40ff-bb7b-f6f6415289ae {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1562.682597] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231751, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1563.046700] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ff5f65-eb7f-402b-b678-83fb031e2e09 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.054914] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a30d26-7534-4db8-9ed6-7d2e403b8cf9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.086298] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bace6b0c-149d-40d7-b68d-7e2428b3f104 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.094334] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21038432-0c4f-457e-bacf-7efb1bd07898 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.110287] env[70013]: DEBUG nova.compute.provider_tree [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1563.182947] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531513} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1563.183585] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore1] 6b329675-3273-4f8a-8e60-5be0d3f8a19e/6b329675-3273-4f8a-8e60-5be0d3f8a19e.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1563.183806] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1563.184074] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-abda1981-63c2-409f-865a-123f82c42fdd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.191297] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1563.191297] env[70013]: value = "task-4231752" [ 1563.191297] env[70013]: _type = "Task" [ 1563.191297] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1563.200504] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231752, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1563.613737] env[70013]: DEBUG nova.scheduler.client.report [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1563.701705] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231752, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.201394} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1563.702113] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1563.702741] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f0a147-efeb-4613-837a-69512ec42142 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.724734] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 6b329675-3273-4f8a-8e60-5be0d3f8a19e/6b329675-3273-4f8a-8e60-5be0d3f8a19e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1563.725071] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c3e63a2-66b1-4516-910f-0f6696d0e852 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.744308] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1563.744308] env[70013]: value = "task-4231753" [ 1563.744308] env[70013]: _type = "Task" [ 1563.744308] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1563.752406] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231753, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1564.118839] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1564.138455] env[70013]: INFO nova.scheduler.client.report [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Deleted allocations for instance 89a7c112-4f31-40ff-bb7b-f6f6415289ae [ 1564.255839] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231753, 'name': ReconfigVM_Task, 'duration_secs': 0.298866} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1564.257046] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 6b329675-3273-4f8a-8e60-5be0d3f8a19e/6b329675-3273-4f8a-8e60-5be0d3f8a19e.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1564.257167] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06b6b67b-2810-4ac4-988c-516e358ef8fc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1564.263809] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1564.263809] env[70013]: value = "task-4231754" [ 1564.263809] env[70013]: _type = "Task" [ 1564.263809] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1564.274564] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231754, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1564.646912] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d6e51f74-8d8d-4f22-a380-52f6049652a7 tempest-AttachVolumeTestJSON-389518749 tempest-AttachVolumeTestJSON-389518749-project-member] Lock "89a7c112-4f31-40ff-bb7b-f6f6415289ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.047s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1564.774008] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231754, 'name': Rename_Task, 'duration_secs': 0.15232} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1564.774297] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1564.774535] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5515c41e-014e-4d3e-89d7-ab591acdf667 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1564.780895] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1564.780895] env[70013]: value = "task-4231755" [ 1564.780895] env[70013]: _type = "Task" [ 1564.780895] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1564.788463] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231755, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1565.291911] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231755, 'name': PowerOnVM_Task} progress is 89%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1565.792837] env[70013]: DEBUG oslo_vmware.api [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231755, 'name': PowerOnVM_Task, 'duration_secs': 0.618952} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1565.792837] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1565.792837] env[70013]: INFO nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1565.793350] env[70013]: DEBUG nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1565.793644] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30f498e-524e-4234-a902-a422d44a052b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1566.314757] env[70013]: INFO nova.compute.manager [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Took 12.45 seconds to build instance. [ 1566.771422] env[70013]: DEBUG nova.compute.manager [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Received event network-changed-2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1566.771616] env[70013]: DEBUG nova.compute.manager [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Refreshing instance network info cache due to event network-changed-2b1ad1e3-7054-4541-8a7d-1987779f9a9a. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1566.771855] env[70013]: DEBUG oslo_concurrency.lockutils [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] Acquiring lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1566.772090] env[70013]: DEBUG oslo_concurrency.lockutils [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] Acquired lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1566.772271] env[70013]: DEBUG nova.network.neutron [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Refreshing network info cache for port 2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1566.817248] env[70013]: DEBUG oslo_concurrency.lockutils [None req-5fa281c0-ed1f-4f14-8559-7821390fae83 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.962s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1567.495281] env[70013]: DEBUG nova.network.neutron [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Updated VIF entry in instance network info cache for port 2b1ad1e3-7054-4541-8a7d-1987779f9a9a. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1567.495645] env[70013]: DEBUG nova.network.neutron [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Updating instance_info_cache with network_info: [{"id": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "address": "fa:16:3e:ae:17:55", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b1ad1e3-70", "ovs_interfaceid": "2b1ad1e3-7054-4541-8a7d-1987779f9a9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1567.998875] env[70013]: DEBUG oslo_concurrency.lockutils [req-0f2e3132-9cef-4b14-8fcb-80005532c2f8 req-491633ba-1832-483c-8c02-017494cbc52f service nova] Releasing lock "refresh_cache-6b329675-3273-4f8a-8e60-5be0d3f8a19e" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1586.064576] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "04f78d9e-cdb5-4173-8b28-3089dc846200" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1586.065103] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1586.567749] env[70013]: DEBUG nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1587.090705] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1587.090987] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1587.092892] env[70013]: INFO nova.compute.claims [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1588.238942] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b612ef45-2bc4-4f6f-a0cf-e85b7b7be84c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.247215] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02cd7983-932f-416c-9108-91aa83067500 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.276779] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7384d6-3f5c-4bc3-b317-2f69c4047daa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.284615] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4917d1-1812-47b4-b38a-8a880107322d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.298008] env[70013]: DEBUG nova.compute.provider_tree [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1588.801442] env[70013]: DEBUG nova.scheduler.client.report [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1589.306837] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.216s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1589.307321] env[70013]: DEBUG nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1589.812982] env[70013]: DEBUG nova.compute.utils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1589.814729] env[70013]: DEBUG nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Allocating IP information in the background. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1589.814729] env[70013]: DEBUG nova.network.neutron [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] allocate_for_instance() {{(pid=70013) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1589.863349] env[70013]: DEBUG nova.policy [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ade87067e834404adc174783ff7b03d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66804721098f48d69e216807291268d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=70013) authorize /opt/stack/nova/nova/policy.py:192}} [ 1590.140472] env[70013]: DEBUG nova.network.neutron [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Successfully created port: 6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1590.318729] env[70013]: DEBUG nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1590.824338] env[70013]: INFO nova.virt.block_device [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Booting with volume 64bab45b-3954-40be-a6b2-40d195e68364 at /dev/sda [ 1590.864041] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3a0ec42-cda7-4fa2-a20e-5fd7c82fc623 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1590.874138] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48961a8-d73e-4c9e-b826-ac832ed826e2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1590.902475] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5180f4db-a892-4bdf-917a-c1d1f32889b0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1590.913034] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93ad953-1fff-46d4-9ba9-47c83eafae54 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1590.939388] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5947b0f-334f-4fd8-b23d-378e6517b274 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1590.946190] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f27dbb-50d3-49c1-a524-f796e05fde0e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1590.960021] env[70013]: DEBUG nova.virt.block_device [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating existing volume attachment record: be0f0fb0-4171-4867-9ef9-056349c8af4b {{(pid=70013) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1591.496522] env[70013]: DEBUG nova.compute.manager [req-c5cf5255-899f-465f-a2c7-3c0574a68d1a req-1d3e3579-309b-4d8f-b94e-7c638f59c65b service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Received event network-vif-plugged-6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1591.496765] env[70013]: DEBUG oslo_concurrency.lockutils [req-c5cf5255-899f-465f-a2c7-3c0574a68d1a req-1d3e3579-309b-4d8f-b94e-7c638f59c65b service nova] Acquiring lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1591.496956] env[70013]: DEBUG oslo_concurrency.lockutils [req-c5cf5255-899f-465f-a2c7-3c0574a68d1a req-1d3e3579-309b-4d8f-b94e-7c638f59c65b service nova] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1591.497171] env[70013]: DEBUG oslo_concurrency.lockutils [req-c5cf5255-899f-465f-a2c7-3c0574a68d1a req-1d3e3579-309b-4d8f-b94e-7c638f59c65b service nova] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1591.497352] env[70013]: DEBUG nova.compute.manager [req-c5cf5255-899f-465f-a2c7-3c0574a68d1a req-1d3e3579-309b-4d8f-b94e-7c638f59c65b service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] No waiting events found dispatching network-vif-plugged-6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) pop_instance_event /opt/stack/nova/nova/compute/manager.py:321}} [ 1591.497520] env[70013]: WARNING nova.compute.manager [req-c5cf5255-899f-465f-a2c7-3c0574a68d1a req-1d3e3579-309b-4d8f-b94e-7c638f59c65b service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Received unexpected event network-vif-plugged-6024653f-4ed6-45dc-8bfe-02867a4a3ca8 for instance with vm_state building and task_state block_device_mapping. [ 1591.581264] env[70013]: DEBUG nova.network.neutron [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Successfully updated port: 6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1592.085242] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1592.085433] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1592.085622] env[70013]: DEBUG nova.network.neutron [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1592.616483] env[70013]: DEBUG nova.network.neutron [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1592.737179] env[70013]: DEBUG nova.network.neutron [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [{"id": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "address": "fa:16:3e:4d:f5:b1", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6024653f-4e", "ovs_interfaceid": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1593.046733] env[70013]: DEBUG nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1593.047424] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1593.047665] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1593.047827] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1593.048041] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1593.048267] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1593.048434] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1593.048668] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1593.048848] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1593.049035] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1593.049229] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1593.049438] env[70013]: DEBUG nova.virt.hardware [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1593.050399] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6beaa2a9-99ba-4b98-bf91-32d079818078 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.059838] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b95331e-d5a6-4a0c-98c9-e6c90a3b0da5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.240100] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1593.240495] env[70013]: DEBUG nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Instance network_info: |[{"id": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "address": "fa:16:3e:4d:f5:b1", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6024653f-4e", "ovs_interfaceid": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1593.240986] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:f5:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6024653f-4ed6-45dc-8bfe-02867a4a3ca8', 'vif_model': 'vmxnet3'}] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1593.248694] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1593.248931] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1593.249188] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4983fa56-efe4-4091-a3ae-569c523b96fa {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.270076] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1593.270076] env[70013]: value = "task-4231767" [ 1593.270076] env[70013]: _type = "Task" [ 1593.270076] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1593.279383] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231767, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1593.530106] env[70013]: DEBUG nova.compute.manager [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Received event network-changed-6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1593.530262] env[70013]: DEBUG nova.compute.manager [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Refreshing instance network info cache due to event network-changed-6024653f-4ed6-45dc-8bfe-02867a4a3ca8. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1593.530506] env[70013]: DEBUG oslo_concurrency.lockutils [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] Acquiring lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1593.530650] env[70013]: DEBUG oslo_concurrency.lockutils [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] Acquired lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1593.530835] env[70013]: DEBUG nova.network.neutron [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Refreshing network info cache for port 6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1593.780170] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231767, 'name': CreateVM_Task, 'duration_secs': 0.309701} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1593.780428] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1593.781019] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'be0f0fb0-4171-4867-9ef9-056349c8af4b', 'guest_format': None, 'device_type': None, 'mount_device': '/dev/sda', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837260', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'name': 'volume-64bab45b-3954-40be-a6b2-40d195e68364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '04f78d9e-cdb5-4173-8b28-3089dc846200', 'attached_at': '', 'detached_at': '', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'serial': '64bab45b-3954-40be-a6b2-40d195e68364'}, 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=70013) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1593.781243] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Root volume attach. Driver type: vmdk {{(pid=70013) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1593.781995] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d067ea58-e316-4f98-8dad-63359474aa16 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.789415] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cf80fb-7c92-4e82-aada-fd174dbfcd21 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.795436] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ea4d5a-d322-4927-b4b1-cae5af264342 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.801435] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-1280b0f3-2b0a-4e8a-b20e-29e1043ec0d3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.809306] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1593.809306] env[70013]: value = "task-4231768" [ 1593.809306] env[70013]: _type = "Task" [ 1593.809306] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1593.816892] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231768, 'name': RelocateVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1594.226919] env[70013]: DEBUG nova.network.neutron [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updated VIF entry in instance network info cache for port 6024653f-4ed6-45dc-8bfe-02867a4a3ca8. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1594.227412] env[70013]: DEBUG nova.network.neutron [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [{"id": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "address": "fa:16:3e:4d:f5:b1", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6024653f-4e", "ovs_interfaceid": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1594.319183] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231768, 'name': RelocateVM_Task, 'duration_secs': 0.027614} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1594.319567] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Volume attach. Driver type: vmdk {{(pid=70013) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1594.319779] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837260', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'name': 'volume-64bab45b-3954-40be-a6b2-40d195e68364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '04f78d9e-cdb5-4173-8b28-3089dc846200', 'attached_at': '', 'detached_at': '', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'serial': '64bab45b-3954-40be-a6b2-40d195e68364'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1594.320613] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4137b5-62b3-4d8b-889d-d8c61d9ca82c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1594.337390] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a2f160-e018-4ab6-92a2-8a818a560837 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1594.358598] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] volume-64bab45b-3954-40be-a6b2-40d195e68364/volume-64bab45b-3954-40be-a6b2-40d195e68364.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1594.358870] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56d36ec0-746e-4f66-a9c4-a046b33b9e7a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1594.380842] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1594.380842] env[70013]: value = "task-4231769" [ 1594.380842] env[70013]: _type = "Task" [ 1594.380842] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1594.389386] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231769, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1594.730176] env[70013]: DEBUG oslo_concurrency.lockutils [req-fa6c1298-0c8d-4e85-9310-eaa5fa7c7653 req-c41f95d3-8277-4222-9fe6-7d3ad751904f service nova] Releasing lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1594.891327] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231769, 'name': ReconfigVM_Task, 'duration_secs': 0.250352} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1594.891698] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfigured VM instance instance-00000075 to attach disk [datastore2] volume-64bab45b-3954-40be-a6b2-40d195e68364/volume-64bab45b-3954-40be-a6b2-40d195e68364.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1594.896279] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c75e8d1-d095-41b5-b337-e02316eedd47 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1594.910637] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1594.910637] env[70013]: value = "task-4231770" [ 1594.910637] env[70013]: _type = "Task" [ 1594.910637] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1594.918647] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231770, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1595.421510] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231770, 'name': ReconfigVM_Task, 'duration_secs': 0.126306} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1595.421828] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837260', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'name': 'volume-64bab45b-3954-40be-a6b2-40d195e68364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '04f78d9e-cdb5-4173-8b28-3089dc846200', 'attached_at': '', 'detached_at': '', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'serial': '64bab45b-3954-40be-a6b2-40d195e68364'} {{(pid=70013) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1595.422405] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c823b637-dc17-4b17-b9ec-ef54a92968cf {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1595.429404] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1595.429404] env[70013]: value = "task-4231771" [ 1595.429404] env[70013]: _type = "Task" [ 1595.429404] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1595.437502] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231771, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1595.940299] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231771, 'name': Rename_Task, 'duration_secs': 0.122468} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1595.940659] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1595.940821] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23b71a87-20a9-4d7c-8ebe-1b2c825b5f27 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1595.947363] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1595.947363] env[70013]: value = "task-4231772" [ 1595.947363] env[70013]: _type = "Task" [ 1595.947363] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1595.956022] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1596.457946] env[70013]: DEBUG oslo_vmware.api [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231772, 'name': PowerOnVM_Task, 'duration_secs': 0.450279} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1596.458267] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1596.458477] env[70013]: INFO nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Took 3.41 seconds to spawn the instance on the hypervisor. [ 1596.458659] env[70013]: DEBUG nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1596.459466] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71143fab-1cba-4fc5-ad1a-bf9dce109e3e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1596.978079] env[70013]: INFO nova.compute.manager [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Took 9.91 seconds to build instance. [ 1597.480097] env[70013]: DEBUG oslo_concurrency.lockutils [None req-1636b817-6a9f-46b9-a537-e5d82b2fa3a1 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.415s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1598.227640] env[70013]: DEBUG nova.compute.manager [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Received event network-changed-09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1598.227977] env[70013]: DEBUG nova.compute.manager [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Refreshing instance network info cache due to event network-changed-09fcfde5-581c-45f5-94e0-d8ed1b20d071. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1598.228072] env[70013]: DEBUG oslo_concurrency.lockutils [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] Acquiring lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1598.228195] env[70013]: DEBUG oslo_concurrency.lockutils [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] Acquired lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1598.228343] env[70013]: DEBUG nova.network.neutron [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Refreshing network info cache for port 09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1598.967528] env[70013]: DEBUG nova.network.neutron [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updated VIF entry in instance network info cache for port 09fcfde5-581c-45f5-94e0-d8ed1b20d071. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1598.967888] env[70013]: DEBUG nova.network.neutron [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updating instance_info_cache with network_info: [{"id": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "address": "fa:16:3e:3e:6c:cf", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09fcfde5-58", "ovs_interfaceid": "09fcfde5-581c-45f5-94e0-d8ed1b20d071", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1599.471385] env[70013]: DEBUG oslo_concurrency.lockutils [req-c050f45a-1ae5-4146-8f03-f878e4f360ed req-d2ae79a0-fa2a-4f29-b1d6-ac8417c28ed7 service nova] Releasing lock "refresh_cache-8d90b797-118d-45cb-9422-7137b7fadbe2" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1599.773705] env[70013]: DEBUG nova.compute.manager [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Stashing vm_state: active {{(pid=70013) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1600.262073] env[70013]: DEBUG nova.compute.manager [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Received event network-changed-6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1600.262383] env[70013]: DEBUG nova.compute.manager [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Refreshing instance network info cache due to event network-changed-6024653f-4ed6-45dc-8bfe-02867a4a3ca8. {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11772}} [ 1600.262697] env[70013]: DEBUG oslo_concurrency.lockutils [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] Acquiring lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1600.262900] env[70013]: DEBUG oslo_concurrency.lockutils [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] Acquired lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1600.263147] env[70013]: DEBUG nova.network.neutron [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Refreshing network info cache for port 6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1600.294023] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1600.294200] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1600.799711] env[70013]: INFO nova.compute.claims [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1600.981460] env[70013]: DEBUG nova.network.neutron [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updated VIF entry in instance network info cache for port 6024653f-4ed6-45dc-8bfe-02867a4a3ca8. {{(pid=70013) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1600.981841] env[70013]: DEBUG nova.network.neutron [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [{"id": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "address": "fa:16:3e:4d:f5:b1", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6024653f-4e", "ovs_interfaceid": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1601.305912] env[70013]: INFO nova.compute.resource_tracker [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating resource usage from migration d15a4aab-e74b-4e2f-976f-55ea9aefa315 [ 1601.448701] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386d0e13-1798-4ad7-8b65-7683fd9ce881 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.456482] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305d189d-5c12-48d9-a154-dfff60fa59b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.486621] env[70013]: DEBUG oslo_concurrency.lockutils [req-bd2469dc-ea68-4db3-bd85-517715ebdd38 req-00a35b32-d03a-4668-8d78-4ac8857e14f6 service nova] Releasing lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1601.487677] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524494a2-cfb7-4526-a981-9131f7a05207 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.495273] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7936fd28-0b71-4c05-b84a-d34de322a106 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.508514] env[70013]: DEBUG nova.compute.provider_tree [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1602.011807] env[70013]: DEBUG nova.scheduler.client.report [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1602.517349] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.223s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1602.517637] env[70013]: INFO nova.compute.manager [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Migrating [ 1603.032883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1603.033339] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1603.033339] env[70013]: DEBUG nova.network.neutron [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1603.745535] env[70013]: DEBUG nova.network.neutron [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [{"id": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "address": "fa:16:3e:4d:f5:b1", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6024653f-4e", "ovs_interfaceid": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1604.248459] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1605.765056] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d318b87-260c-45fa-b683-49feb2b2e4fe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.784431] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance '04f78d9e-cdb5-4173-8b28-3089dc846200' progress to 0 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1605.878877] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1606.290880] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1606.291166] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e401a67-3a04-4307-a9bb-4611bcd65e11 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.299442] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1606.299442] env[70013]: value = "task-4231773" [ 1606.299442] env[70013]: _type = "Task" [ 1606.299442] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1606.309231] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1606.382603] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1606.382906] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1606.383095] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1606.383259] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1606.384402] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7b986c-5d96-4c1a-8b19-f53f8db166a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.396355] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0027ca8e-0652-42a2-af2e-58b8cb872c66 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.416641] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0fed8f9-388e-43fa-b512-c08b10ad174c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.424682] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cfaace-f26a-4641-a34e-050bf463a2f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.456142] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180264MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1606.456309] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1606.456596] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1606.809330] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231773, 'name': PowerOffVM_Task, 'duration_secs': 0.238625} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1606.809749] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1606.809792] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance '04f78d9e-cdb5-4173-8b28-3089dc846200' progress to 17 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1607.316627] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1607.316880] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1607.317099] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1607.317244] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1607.317434] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1607.317620] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1607.317842] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1607.318012] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1607.318203] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1607.318436] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1607.318640] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1607.324387] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e3fe245-88a8-4fa5-a4b4-ebaf29650129 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1607.341293] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1607.341293] env[70013]: value = "task-4231774" [ 1607.341293] env[70013]: _type = "Task" [ 1607.341293] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1607.349985] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231774, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1607.467732] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Applying migration context for instance 04f78d9e-cdb5-4173-8b28-3089dc846200 as it has an incoming, in-progress migration d15a4aab-e74b-4e2f-976f-55ea9aefa315. Migration status is migrating {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1607.469350] env[70013]: INFO nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating resource usage from migration d15a4aab-e74b-4e2f-976f-55ea9aefa315 [ 1607.490784] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.490923] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491026] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491205] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491270] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491327] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491480] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491538] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491655] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8d90b797-118d-45cb-9422-7137b7fadbe2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491845] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 6b329675-3273-4f8a-8e60-5be0d3f8a19e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.491933] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Migration d15a4aab-e74b-4e2f-976f-55ea9aefa315 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1742}} [ 1607.491996] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 04f78d9e-cdb5-4173-8b28-3089dc846200 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1607.492473] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1607.492473] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2944MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1607.646946] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9e2a07-4818-4504-ba05-b60e753a0e40 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1607.657296] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4814efad-afc6-4aca-a3e0-b3283d7e529d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1607.694256] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b29879-a7a9-42db-8d40-8dbc1a236eae {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1607.702277] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e5b417-18c2-4b4f-a263-d7bbb6ddf6e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1607.716020] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1607.852247] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231774, 'name': ReconfigVM_Task, 'duration_secs': 0.180533} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1607.852623] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance '04f78d9e-cdb5-4173-8b28-3089dc846200' progress to 33 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1608.219496] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1608.359467] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1608.359765] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1608.359939] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1608.360145] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1608.360295] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1608.360448] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1608.360743] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1608.360968] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1608.361176] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1608.361347] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1608.361522] env[70013]: DEBUG nova.virt.hardware [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1608.367122] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1608.368265] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3e09e42-525e-4019-bb9c-408b210c6184 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1608.388015] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1608.388015] env[70013]: value = "task-4231775" [ 1608.388015] env[70013]: _type = "Task" [ 1608.388015] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1608.396324] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231775, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1608.724343] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1608.724907] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.268s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1608.897751] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231775, 'name': ReconfigVM_Task, 'duration_secs': 0.15489} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1608.898178] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1608.898798] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32492a08-0229-4b38-9e49-67f3ba60ea85 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1608.920720] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] volume-64bab45b-3954-40be-a6b2-40d195e68364/volume-64bab45b-3954-40be-a6b2-40d195e68364.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1608.920979] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-824662bc-0e0b-48f4-be78-f2b6e57ccad1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1608.939110] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1608.939110] env[70013]: value = "task-4231776" [ 1608.939110] env[70013]: _type = "Task" [ 1608.939110] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1608.946980] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231776, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1609.186142] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1609.186338] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1609.186507] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1609.186667] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1609.186838] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1609.186995] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1609.187157] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1609.449402] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231776, 'name': ReconfigVM_Task, 'duration_secs': 0.498091} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1609.449642] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfigured VM instance instance-00000075 to attach disk [datastore2] volume-64bab45b-3954-40be-a6b2-40d195e68364/volume-64bab45b-3954-40be-a6b2-40d195e68364.vmdk or device None with type thin {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1609.449898] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance '04f78d9e-cdb5-4173-8b28-3089dc846200' progress to 50 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1609.957146] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c402f6d-772e-43c9-9a5a-a56c76e82915 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.981095] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0feadb1-352f-431c-96f7-3873e7addebd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.999717] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance '04f78d9e-cdb5-4173-8b28-3089dc846200' progress to 67 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1611.678370] env[70013]: DEBUG nova.network.neutron [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Port 6024653f-4ed6-45dc-8bfe-02867a4a3ca8 binding to destination host cpu-1 is already ACTIVE {{(pid=70013) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1612.699961] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1612.700397] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1612.700397] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1613.335111] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1613.737651] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1613.737883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1613.738027] env[70013]: DEBUG nova.network.neutron [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1614.339433] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1614.473521] env[70013]: DEBUG nova.network.neutron [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [{"id": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "address": "fa:16:3e:4d:f5:b1", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6024653f-4e", "ovs_interfaceid": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1614.977063] env[70013]: DEBUG oslo_concurrency.lockutils [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1615.487835] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3404b2dc-9f61-46ff-80d5-f9eca9ce7801 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1615.495480] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1acd323d-70c4-458f-ba71-0b07ceee335a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.598475] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3033e801-c16a-411d-917c-e1585d85fed6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.618371] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eebe494-e2e5-48e2-bd51-c707f28c3778 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.625797] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance '04f78d9e-cdb5-4173-8b28-3089dc846200' progress to 83 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1617.131937] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1617.132286] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2297920-c1c4-494e-bc9f-85a1a0732421 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1617.140175] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1617.140175] env[70013]: value = "task-4231777" [ 1617.140175] env[70013]: _type = "Task" [ 1617.140175] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1617.148760] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1617.650436] env[70013]: DEBUG oslo_vmware.api [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231777, 'name': PowerOnVM_Task, 'duration_secs': 0.403099} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1617.650780] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1617.650880] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-4df4ad8e-346b-4c3d-9789-dba0b71e6524 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance '04f78d9e-cdb5-4173-8b28-3089dc846200' progress to 100 {{(pid=70013) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1619.470530] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "04f78d9e-cdb5-4173-8b28-3089dc846200" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1619.470984] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1619.470984] env[70013]: DEBUG nova.compute.manager [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Going to confirm migration 5 {{(pid=70013) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1620.053911] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1620.054133] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquired lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1620.054316] env[70013]: DEBUG nova.network.neutron [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1620.054569] env[70013]: DEBUG nova.objects.instance [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'info_cache' on Instance uuid 04f78d9e-cdb5-4173-8b28-3089dc846200 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1621.293222] env[70013]: DEBUG nova.network.neutron [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [{"id": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "address": "fa:16:3e:4d:f5:b1", "network": {"id": "b2c2a219-c3ac-4bf1-a554-19513dfe6bf5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-287271651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66804721098f48d69e216807291268d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6024653f-4e", "ovs_interfaceid": "6024653f-4ed6-45dc-8bfe-02867a4a3ca8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1621.796081] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Releasing lock "refresh_cache-04f78d9e-cdb5-4173-8b28-3089dc846200" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1621.796376] env[70013]: DEBUG nova.objects.instance [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'migration_context' on Instance uuid 04f78d9e-cdb5-4173-8b28-3089dc846200 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1622.299772] env[70013]: DEBUG nova.objects.base [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Object Instance<04f78d9e-cdb5-4173-8b28-3089dc846200> lazy-loaded attributes: info_cache,migration_context {{(pid=70013) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1622.300775] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419948da-80f7-4167-8f92-0e5c872b7a33 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1622.321161] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f68459d-c10a-46f7-8ea1-caf227f4cdd8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1622.326666] env[70013]: DEBUG oslo_vmware.api [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1622.326666] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3c9c2-32db-d885-f008-d1adf58c22c3" [ 1622.326666] env[70013]: _type = "Task" [ 1622.326666] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1622.336321] env[70013]: DEBUG oslo_vmware.api [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3c9c2-32db-d885-f008-d1adf58c22c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1622.837680] env[70013]: DEBUG oslo_vmware.api [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3c9c2-32db-d885-f008-d1adf58c22c3, 'name': SearchDatastore_Task, 'duration_secs': 0.01073} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1622.837973] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1622.838238] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1623.489956] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95ef147-f5eb-4ffd-82fb-4c6118a1bb9b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.497963] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d669648-6e50-4a1d-be44-3c63c3c4d6cb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.527993] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790cb890-eeb6-40ba-943a-29b64cedc761 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.535860] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c4dc30-09d0-4d5d-bb5e-58f324fd9939 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.549217] env[70013]: DEBUG nova.compute.provider_tree [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1624.052916] env[70013]: DEBUG nova.scheduler.client.report [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1625.064694] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.226s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1625.622719] env[70013]: INFO nova.scheduler.client.report [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted allocation for migration d15a4aab-e74b-4e2f-976f-55ea9aefa315 [ 1625.931279] env[70013]: INFO nova.compute.manager [None req-d0c19df2-2a7e-48e1-93e5-c3ba5cd1dc5e tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Get console output [ 1625.931658] env[70013]: WARNING nova.virt.vmwareapi.driver [None req-d0c19df2-2a7e-48e1-93e5-c3ba5cd1dc5e tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] The console log is missing. Check your VSPC configuration [ 1626.128516] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6daba7ac-0a1f-4b2f-8828-347217ea40a6 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.658s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1652.972868] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "04f78d9e-cdb5-4173-8b28-3089dc846200" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1652.973284] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1652.973325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1652.973515] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1652.973685] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1652.977194] env[70013]: INFO nova.compute.manager [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Terminating instance [ 1653.481940] env[70013]: DEBUG nova.compute.manager [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1653.482167] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1653.482483] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e969cad7-0b6c-4917-85a9-d19f429ae067 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.491175] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1653.491175] env[70013]: value = "task-4231778" [ 1653.491175] env[70013]: _type = "Task" [ 1653.491175] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1653.500982] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1654.002286] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231778, 'name': PowerOffVM_Task, 'duration_secs': 0.174968} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1654.002631] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1654.002779] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Volume detach. Driver type: vmdk {{(pid=70013) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1654.002978] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837260', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'name': 'volume-64bab45b-3954-40be-a6b2-40d195e68364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '04f78d9e-cdb5-4173-8b28-3089dc846200', 'attached_at': '2025-05-19T11:48:34.000000', 'detached_at': '', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'serial': '64bab45b-3954-40be-a6b2-40d195e68364'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1654.003806] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790da135-ed18-4ced-bc61-30578920de46 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.028655] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82edec38-e0bd-4fc5-b5f4-a4b5d9b9591f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.038379] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cfacfd3-c32e-49cc-b6e3-bb32339565e7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.057732] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6ec92d-00e5-49bf-938b-0d180b716e51 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.073380] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] The volume has not been displaced from its original location: [datastore2] volume-64bab45b-3954-40be-a6b2-40d195e68364/volume-64bab45b-3954-40be-a6b2-40d195e68364.vmdk. No consolidation needed. {{(pid=70013) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1654.078818] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1654.079125] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f376d334-0097-4a49-bb2a-6d9d4c66194a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.097976] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1654.097976] env[70013]: value = "task-4231779" [ 1654.097976] env[70013]: _type = "Task" [ 1654.097976] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1654.106260] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231779, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1654.608432] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231779, 'name': ReconfigVM_Task, 'duration_secs': 0.157555} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1654.608721] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=70013) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1654.613305] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8763548-e6ff-4199-b213-6d9ab82cc794 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.628952] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1654.628952] env[70013]: value = "task-4231780" [ 1654.628952] env[70013]: _type = "Task" [ 1654.628952] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1654.637399] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231780, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1655.138898] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231780, 'name': ReconfigVM_Task, 'duration_secs': 0.139855} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1655.140062] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-837260', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'name': 'volume-64bab45b-3954-40be-a6b2-40d195e68364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '04f78d9e-cdb5-4173-8b28-3089dc846200', 'attached_at': '2025-05-19T11:48:34.000000', 'detached_at': '', 'volume_id': '64bab45b-3954-40be-a6b2-40d195e68364', 'serial': '64bab45b-3954-40be-a6b2-40d195e68364'} {{(pid=70013) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1655.140202] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1655.140908] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9e5172-e0b4-48b2-8c4b-d2a8c2a6ccc1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1655.147532] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1655.147767] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06a69c9a-6f9d-449f-8661-c243f1e3ae66 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1655.212776] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1655.212991] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1655.213181] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleting the datastore file [datastore2] 04f78d9e-cdb5-4173-8b28-3089dc846200 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1655.213460] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98d76d23-fd8b-46ee-a209-9d26f4519c1e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1655.220393] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1655.220393] env[70013]: value = "task-4231782" [ 1655.220393] env[70013]: _type = "Task" [ 1655.220393] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1655.228601] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231782, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1655.731035] env[70013]: DEBUG oslo_vmware.api [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231782, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246684} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1655.731035] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1655.731035] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1655.731292] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1655.731328] env[70013]: INFO nova.compute.manager [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Took 2.25 seconds to destroy the instance on the hypervisor. [ 1655.731555] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1655.731772] env[70013]: DEBUG nova.compute.manager [-] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1655.731872] env[70013]: DEBUG nova.network.neutron [-] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1656.226587] env[70013]: DEBUG nova.compute.manager [req-50bcb1b1-30e1-423e-82e8-283e8e2bccd1 req-98cfad98-4d78-419e-8a3d-91bf73ad650b service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Received event network-vif-deleted-6024653f-4ed6-45dc-8bfe-02867a4a3ca8 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1656.226872] env[70013]: INFO nova.compute.manager [req-50bcb1b1-30e1-423e-82e8-283e8e2bccd1 req-98cfad98-4d78-419e-8a3d-91bf73ad650b service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Neutron deleted interface 6024653f-4ed6-45dc-8bfe-02867a4a3ca8; detaching it from the instance and deleting it from the info cache [ 1656.227110] env[70013]: DEBUG nova.network.neutron [req-50bcb1b1-30e1-423e-82e8-283e8e2bccd1 req-98cfad98-4d78-419e-8a3d-91bf73ad650b service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1656.676967] env[70013]: DEBUG nova.network.neutron [-] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1656.730034] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f4da135-2086-42c0-adb7-210d4b3f3340 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1656.739695] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33f4ce1-6765-4205-87ac-87811319a033 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1656.765292] env[70013]: DEBUG nova.compute.manager [req-50bcb1b1-30e1-423e-82e8-283e8e2bccd1 req-98cfad98-4d78-419e-8a3d-91bf73ad650b service nova] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Detach interface failed, port_id=6024653f-4ed6-45dc-8bfe-02867a4a3ca8, reason: Instance 04f78d9e-cdb5-4173-8b28-3089dc846200 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1657.179526] env[70013]: INFO nova.compute.manager [-] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Took 1.45 seconds to deallocate network for instance. [ 1657.728044] env[70013]: INFO nova.compute.manager [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Took 0.55 seconds to detach 1 volumes for instance. [ 1657.730302] env[70013]: DEBUG nova.compute.manager [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 04f78d9e-cdb5-4173-8b28-3089dc846200] Deleting volume: 64bab45b-3954-40be-a6b2-40d195e68364 {{(pid=70013) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 1658.268022] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1658.268296] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1658.268478] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1658.291176] env[70013]: INFO nova.scheduler.client.report [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted allocations for instance 04f78d9e-cdb5-4173-8b28-3089dc846200 [ 1658.799384] env[70013]: DEBUG oslo_concurrency.lockutils [None req-750ae828-0833-4429-be19-258446be4fb0 tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "04f78d9e-cdb5-4173-8b28-3089dc846200" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.826s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1659.663617] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1659.663842] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1659.664072] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1659.664324] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1659.664447] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1659.666723] env[70013]: INFO nova.compute.manager [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Terminating instance [ 1660.170504] env[70013]: DEBUG nova.compute.manager [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1660.170873] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1660.171600] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dd0a32-d72a-4869-9b40-1d613eeae255 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1660.179798] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1660.180060] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82b65016-f079-4b92-80fb-f0acaf8ee157 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1660.186457] env[70013]: DEBUG oslo_vmware.api [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1660.186457] env[70013]: value = "task-4231784" [ 1660.186457] env[70013]: _type = "Task" [ 1660.186457] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1660.194781] env[70013]: DEBUG oslo_vmware.api [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1660.697357] env[70013]: DEBUG oslo_vmware.api [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231784, 'name': PowerOffVM_Task, 'duration_secs': 0.190536} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1660.697628] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1660.697831] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1660.698110] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee9ec86b-b7f4-459e-90e9-c6391deb4319 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1660.767602] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1660.767903] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Deleting contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1660.768069] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleting the datastore file [datastore1] 6b329675-3273-4f8a-8e60-5be0d3f8a19e {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1660.768340] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d0326fb9-9370-4c12-a0c5-9247df5cfdd3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1660.774715] env[70013]: DEBUG oslo_vmware.api [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1660.774715] env[70013]: value = "task-4231786" [ 1660.774715] env[70013]: _type = "Task" [ 1660.774715] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1660.784888] env[70013]: DEBUG oslo_vmware.api [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231786, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1661.285173] env[70013]: DEBUG oslo_vmware.api [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231786, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141309} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1661.285557] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1661.285599] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Deleted contents of the VM from datastore datastore1 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1661.285781] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1661.285961] env[70013]: INFO nova.compute.manager [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1661.286231] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1661.286433] env[70013]: DEBUG nova.compute.manager [-] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1661.286528] env[70013]: DEBUG nova.network.neutron [-] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1661.541231] env[70013]: DEBUG nova.compute.manager [req-8461f722-0192-4b9e-ad50-fe671686aeb7 req-9d6b6116-17de-40a4-84bf-f5e5d86addeb service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Received event network-vif-deleted-2b1ad1e3-7054-4541-8a7d-1987779f9a9a {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1661.541393] env[70013]: INFO nova.compute.manager [req-8461f722-0192-4b9e-ad50-fe671686aeb7 req-9d6b6116-17de-40a4-84bf-f5e5d86addeb service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Neutron deleted interface 2b1ad1e3-7054-4541-8a7d-1987779f9a9a; detaching it from the instance and deleting it from the info cache [ 1661.541525] env[70013]: DEBUG nova.network.neutron [req-8461f722-0192-4b9e-ad50-fe671686aeb7 req-9d6b6116-17de-40a4-84bf-f5e5d86addeb service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1662.023579] env[70013]: DEBUG nova.network.neutron [-] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1662.043870] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c0d5e89-ae68-41fa-ba74-90b35a337121 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1662.054933] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df86432-9b3d-4635-9251-915284b0c0ed {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1662.079917] env[70013]: DEBUG nova.compute.manager [req-8461f722-0192-4b9e-ad50-fe671686aeb7 req-9d6b6116-17de-40a4-84bf-f5e5d86addeb service nova] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Detach interface failed, port_id=2b1ad1e3-7054-4541-8a7d-1987779f9a9a, reason: Instance 6b329675-3273-4f8a-8e60-5be0d3f8a19e could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1662.526538] env[70013]: INFO nova.compute.manager [-] [instance: 6b329675-3273-4f8a-8e60-5be0d3f8a19e] Took 1.24 seconds to deallocate network for instance. [ 1663.033486] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1663.033772] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1663.034014] env[70013]: DEBUG nova.objects.instance [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'resources' on Instance uuid 6b329675-3273-4f8a-8e60-5be0d3f8a19e {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1663.659517] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9fc04c-0a6f-441b-bb60-c3a98018bc3c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.667445] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce036fa1-0e0f-43a3-896d-66407d5d2784 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.698936] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c21f2a-8242-4d18-9a35-1bf44cc843df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.706869] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5a38d4-d490-4223-a9c5-356ddf852a83 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.721755] env[70013]: DEBUG nova.compute.provider_tree [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1664.225024] env[70013]: DEBUG nova.scheduler.client.report [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1664.339858] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1664.730330] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1664.750558] env[70013]: INFO nova.scheduler.client.report [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted allocations for instance 6b329675-3273-4f8a-8e60-5be0d3f8a19e [ 1664.842782] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1664.843046] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1664.843223] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1664.843378] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1664.844270] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725c0a21-eb7a-42ff-a90f-ff056cc99cf0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.852604] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab436563-53fd-477e-a4b6-ab46365b6247 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.866921] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0466504e-ece8-481d-b7e2-79bf302108e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.873531] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be0e9e1-da19-4f0e-98d1-a8476315c035 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.902930] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180000MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1664.903129] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1664.903300] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1665.261574] env[70013]: DEBUG oslo_concurrency.lockutils [None req-7f79f1d7-7c44-41e9-9074-80310862701a tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "6b329675-3273-4f8a-8e60-5be0d3f8a19e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.597s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1665.929799] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930117] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930117] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930243] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930341] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930454] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930566] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930681] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930795] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 8d90b797-118d-45cb-9422-7137b7fadbe2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1665.930988] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1665.931142] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1665.947031] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing inventories for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1665.961226] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Updating ProviderTree inventory for provider dc0a4c9d-b611-453e-a900-41e280c29b95 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1665.961425] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Updating inventory in ProviderTree for provider dc0a4c9d-b611-453e-a900-41e280c29b95 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1665.972952] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing aggregate associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, aggregates: None {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1665.993259] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Refreshing trait associations for resource provider dc0a4c9d-b611-453e-a900-41e280c29b95, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=70013) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1666.105610] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86b3ba1-daaf-4a9a-9591-d1c8e5b84026 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1666.113366] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc888d14-5071-4b62-9d2a-b9d11d5890eb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1666.143038] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdaba4c5-e3b3-4da6-8a36-c8b733b13b69 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1666.150957] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261e81b1-33c1-4661-b822-c5e6c055984d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1666.165610] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1666.669366] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1667.174227] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1667.174643] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.271s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1669.414515] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "8d90b797-118d-45cb-9422-7137b7fadbe2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1669.414883] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1669.415419] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "8d90b797-118d-45cb-9422-7137b7fadbe2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1669.415642] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1669.415823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1669.418155] env[70013]: INFO nova.compute.manager [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Terminating instance [ 1669.921942] env[70013]: DEBUG nova.compute.manager [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1669.922251] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1669.923196] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf09f58-217b-4262-9d56-836276afc61b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1669.931720] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1669.932017] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ddc29042-dd01-43fb-b598-abbcac52aabd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1669.938856] env[70013]: DEBUG oslo_vmware.api [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1669.938856] env[70013]: value = "task-4231788" [ 1669.938856] env[70013]: _type = "Task" [ 1669.938856] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1669.949080] env[70013]: DEBUG oslo_vmware.api [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231788, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1670.174737] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1670.174953] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1670.175128] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1670.175268] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1670.175418] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1670.175581] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1670.175772] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1670.450124] env[70013]: DEBUG oslo_vmware.api [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231788, 'name': PowerOffVM_Task, 'duration_secs': 0.178617} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1670.450486] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1670.450621] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1670.450813] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eade7ffc-be69-4171-9071-0008d85d0de9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1670.514479] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1670.514749] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1670.514876] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleting the datastore file [datastore2] 8d90b797-118d-45cb-9422-7137b7fadbe2 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1670.515175] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d02c6292-e2e8-4042-ae87-3167b1c55e94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1670.523212] env[70013]: DEBUG oslo_vmware.api [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for the task: (returnval){ [ 1670.523212] env[70013]: value = "task-4231790" [ 1670.523212] env[70013]: _type = "Task" [ 1670.523212] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1670.531229] env[70013]: DEBUG oslo_vmware.api [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231790, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1671.032942] env[70013]: DEBUG oslo_vmware.api [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Task: {'id': task-4231790, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133034} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1671.033217] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1671.033397] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1671.033578] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1671.033759] env[70013]: INFO nova.compute.manager [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1671.034022] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1671.034221] env[70013]: DEBUG nova.compute.manager [-] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1671.034319] env[70013]: DEBUG nova.network.neutron [-] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1671.291956] env[70013]: DEBUG nova.compute.manager [req-385a5dbd-ba6b-4268-89f3-44e22b64d0d6 req-af2d319d-3144-4756-8dc6-5cf472ab6965 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Received event network-vif-deleted-09fcfde5-581c-45f5-94e0-d8ed1b20d071 {{(pid=70013) external_instance_event /opt/stack/nova/nova/compute/manager.py:11767}} [ 1671.292210] env[70013]: INFO nova.compute.manager [req-385a5dbd-ba6b-4268-89f3-44e22b64d0d6 req-af2d319d-3144-4756-8dc6-5cf472ab6965 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Neutron deleted interface 09fcfde5-581c-45f5-94e0-d8ed1b20d071; detaching it from the instance and deleting it from the info cache [ 1671.292397] env[70013]: DEBUG nova.network.neutron [req-385a5dbd-ba6b-4268-89f3-44e22b64d0d6 req-af2d319d-3144-4756-8dc6-5cf472ab6965 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1671.773635] env[70013]: DEBUG nova.network.neutron [-] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1671.795109] env[70013]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-96be7793-0fd5-4c62-9b8e-8c03fc0ebe1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1671.806889] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1a3914-a115-4b6f-a492-d383f7315d7f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1671.831810] env[70013]: DEBUG nova.compute.manager [req-385a5dbd-ba6b-4268-89f3-44e22b64d0d6 req-af2d319d-3144-4756-8dc6-5cf472ab6965 service nova] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Detach interface failed, port_id=09fcfde5-581c-45f5-94e0-d8ed1b20d071, reason: Instance 8d90b797-118d-45cb-9422-7137b7fadbe2 could not be found. {{(pid=70013) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11601}} [ 1672.278067] env[70013]: INFO nova.compute.manager [-] [instance: 8d90b797-118d-45cb-9422-7137b7fadbe2] Took 1.24 seconds to deallocate network for instance. [ 1672.785507] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1672.785903] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1672.786094] env[70013]: DEBUG nova.objects.instance [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lazy-loading 'resources' on Instance uuid 8d90b797-118d-45cb-9422-7137b7fadbe2 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1673.411136] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2b7a9a-cb80-4fe2-8419-3a957aa53d6a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1673.419090] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3c4470-4606-419c-91cb-ee686d951d9e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1673.448241] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800b3f01-e9c1-4f83-a521-2dd5e7072006 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1673.455474] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4af1ff-3689-47c7-8078-636e71a80736 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1673.470201] env[70013]: DEBUG nova.compute.provider_tree [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1673.973557] env[70013]: DEBUG nova.scheduler.client.report [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1674.478914] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1674.497066] env[70013]: INFO nova.scheduler.client.report [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Deleted allocations for instance 8d90b797-118d-45cb-9422-7137b7fadbe2 [ 1675.004441] env[70013]: DEBUG oslo_concurrency.lockutils [None req-208081ed-4ba2-4fdd-ad57-23bf3479ce2b tempest-ServerActionsTestOtherA-1491726041 tempest-ServerActionsTestOtherA-1491726041-project-member] Lock "8d90b797-118d-45cb-9422-7137b7fadbe2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.589s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1675.340880] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1678.574648] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "538d6d33-5134-4574-9c28-31215aa6df60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1678.574648] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "538d6d33-5134-4574-9c28-31215aa6df60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1678.741827] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "f46c81c7-096d-4ac4-849c-608865ff6a39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1678.742073] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "f46c81c7-096d-4ac4-849c-608865ff6a39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1679.076566] env[70013]: DEBUG nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1679.244733] env[70013]: DEBUG nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1679.601489] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1679.601782] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1679.603616] env[70013]: INFO nova.compute.claims [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1679.763359] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1680.738109] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281afb08-1849-47c0-a074-575f6f411071 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.746246] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d58ade0-0860-4375-ae78-88db198cb58c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.776900] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f741b3-04cb-4b63-83b9-c19da6be6af6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.784661] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38486ca-78e8-4895-af28-b494eb45f275 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.798358] env[70013]: DEBUG nova.compute.provider_tree [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1681.302089] env[70013]: DEBUG nova.scheduler.client.report [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1681.807017] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.205s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1681.807616] env[70013]: DEBUG nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1681.810343] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.047s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1681.811715] env[70013]: INFO nova.compute.claims [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1682.316491] env[70013]: DEBUG nova.compute.utils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1682.319937] env[70013]: DEBUG nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1682.821551] env[70013]: DEBUG nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1682.951068] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a17f420-cc65-442b-ab1c-7637ea1f02a4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1682.959344] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8e6463-c5aa-495d-baa4-3bfd711af1b5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1682.989771] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70066f6d-1b59-4354-925b-fee20a8ec8da {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1682.997518] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15012bb-bafb-4cd9-8737-681a699ef98d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.011069] env[70013]: DEBUG nova.compute.provider_tree [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1683.514042] env[70013]: DEBUG nova.scheduler.client.report [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1683.836129] env[70013]: DEBUG nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1683.863960] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1683.864241] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1683.864402] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1683.864587] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1683.864735] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1683.864884] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1683.865116] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1683.865285] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1683.865455] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1683.865622] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1683.865806] env[70013]: DEBUG nova.virt.hardware [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1683.866689] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb8b549-6d2e-49f8-a8b6-22246b898b1c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.874966] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8099f3d-299d-40fd-8cb9-e54b87827262 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.889995] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1683.895669] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Creating folder: Project (a084ca09083c454299c7eb8c62b37495). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1683.895958] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c1937aa-eeb6-4ae6-8acb-47e3cffa9567 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.909365] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Created folder: Project (a084ca09083c454299c7eb8c62b37495) in parent group-v836999. [ 1683.909554] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Creating folder: Instances. Parent ref: group-v837262. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1683.909790] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c96c66b6-8fde-4a3b-a978-13419973ea1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.918858] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Created folder: Instances in parent group-v837262. [ 1683.919117] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1683.919313] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1683.919511] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db49b265-5e21-4131-9e65-38b7247c0ff2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.935310] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1683.935310] env[70013]: value = "task-4231793" [ 1683.935310] env[70013]: _type = "Task" [ 1683.935310] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1683.942855] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231793, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1684.019019] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1684.019705] env[70013]: DEBUG nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1684.445801] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231793, 'name': CreateVM_Task, 'duration_secs': 0.255733} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1684.446069] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1684.446540] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1684.446709] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1684.447078] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1684.447341] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1dbb06c-2dd0-4ca0-ab6f-991d05fc1461 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1684.451993] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1684.451993] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5232d609-0b9e-df6a-47e8-e48a958acbf2" [ 1684.451993] env[70013]: _type = "Task" [ 1684.451993] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1684.459749] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5232d609-0b9e-df6a-47e8-e48a958acbf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1684.525530] env[70013]: DEBUG nova.compute.utils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1684.527090] env[70013]: DEBUG nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1684.963301] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5232d609-0b9e-df6a-47e8-e48a958acbf2, 'name': SearchDatastore_Task, 'duration_secs': 0.013418} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1684.963701] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1684.963833] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1684.964088] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1684.964293] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1684.964491] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1684.964762] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2aa1f2c5-4913-4031-aad3-e99df1131e59 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1684.973799] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1684.973977] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1684.974713] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a8c38a7-da13-46b5-8af4-5bd2eafcb15a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1684.980216] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1684.980216] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525e3d4a-2a31-ac7a-6b74-b478c2262bcb" [ 1684.980216] env[70013]: _type = "Task" [ 1684.980216] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1684.988186] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525e3d4a-2a31-ac7a-6b74-b478c2262bcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1685.028949] env[70013]: DEBUG nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1685.490784] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525e3d4a-2a31-ac7a-6b74-b478c2262bcb, 'name': SearchDatastore_Task, 'duration_secs': 0.009581} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1685.491592] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2a071ef-58bd-41b2-b874-8b07e27ab529 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1685.496872] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1685.496872] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c22e2-118b-78e1-9316-2f9a96144236" [ 1685.496872] env[70013]: _type = "Task" [ 1685.496872] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1685.505162] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c22e2-118b-78e1-9316-2f9a96144236, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1686.008052] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525c22e2-118b-78e1-9316-2f9a96144236, 'name': SearchDatastore_Task, 'duration_secs': 0.01105} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1686.008459] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1686.008636] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 538d6d33-5134-4574-9c28-31215aa6df60/538d6d33-5134-4574-9c28-31215aa6df60.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1686.008936] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd99c573-0a05-4faa-8f90-39635acfbba8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1686.017679] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1686.017679] env[70013]: value = "task-4231794" [ 1686.017679] env[70013]: _type = "Task" [ 1686.017679] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1686.025820] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1686.038171] env[70013]: DEBUG nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1686.059857] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1686.060236] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1686.060368] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1686.060616] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1686.060832] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1686.061076] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1686.061376] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1686.061567] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1686.061793] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1686.062050] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1686.062300] env[70013]: DEBUG nova.virt.hardware [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1686.063335] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a42975-690b-4e98-95dc-c797444fa6f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1686.072282] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e4b7ff-19b4-442f-94ce-005ba74f26cc {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1686.089137] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1686.095638] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1686.095988] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1686.096264] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-92c630e7-c69c-45b4-9935-df6f07558b1a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1686.115496] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1686.115496] env[70013]: value = "task-4231795" [ 1686.115496] env[70013]: _type = "Task" [ 1686.115496] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1686.127654] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231795, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1686.528198] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456751} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1686.528473] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] 538d6d33-5134-4574-9c28-31215aa6df60/538d6d33-5134-4574-9c28-31215aa6df60.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1686.528692] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1686.528961] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-190ff577-f56e-4fc2-9d6e-89d04e77eebd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1686.536445] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1686.536445] env[70013]: value = "task-4231796" [ 1686.536445] env[70013]: _type = "Task" [ 1686.536445] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1686.545051] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231796, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1686.625584] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231795, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1687.045727] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231796, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076489} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1687.046160] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1687.046882] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cb8c65-080d-4e89-a196-5431d27d7f60 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.066742] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 538d6d33-5134-4574-9c28-31215aa6df60/538d6d33-5134-4574-9c28-31215aa6df60.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1687.067031] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40c7f999-577b-4047-8617-b7d4f21217e4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.087983] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1687.087983] env[70013]: value = "task-4231797" [ 1687.087983] env[70013]: _type = "Task" [ 1687.087983] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1687.096673] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231797, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1687.125302] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231795, 'name': CreateVM_Task, 'duration_secs': 0.538939} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1687.125565] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1687.125935] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1687.126136] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1687.126480] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1687.126750] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-111493e4-f1c9-499e-9d7b-ad30c8c4cb5b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.131740] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1687.131740] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1a26d-d612-c688-4e4e-a7ecabc7f3b2" [ 1687.131740] env[70013]: _type = "Task" [ 1687.131740] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1687.139934] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1a26d-d612-c688-4e4e-a7ecabc7f3b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1687.598567] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231797, 'name': ReconfigVM_Task, 'duration_secs': 0.272277} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1687.598850] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 538d6d33-5134-4574-9c28-31215aa6df60/538d6d33-5134-4574-9c28-31215aa6df60.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1687.599493] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a73660d7-c089-41d6-8ba7-57b0f9f4a195 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.606931] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1687.606931] env[70013]: value = "task-4231798" [ 1687.606931] env[70013]: _type = "Task" [ 1687.606931] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1687.617144] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231798, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1687.643615] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1a26d-d612-c688-4e4e-a7ecabc7f3b2, 'name': SearchDatastore_Task, 'duration_secs': 0.010059} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1687.643870] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1687.644103] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1687.644347] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1687.644497] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1687.644679] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1687.644956] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2eaabe8-e561-438b-8115-e3b14123fabe {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.654669] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1687.654838] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1687.655593] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa869ce9-2b06-4e69-bc3a-93c32d3102c8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.661557] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1687.661557] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]5221a0f5-c318-03d4-c699-23f8cf062cad" [ 1687.661557] env[70013]: _type = "Task" [ 1687.661557] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1687.670862] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5221a0f5-c318-03d4-c699-23f8cf062cad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1688.117941] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231798, 'name': Rename_Task, 'duration_secs': 0.132752} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1688.118341] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1688.118502] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2ea24ff-2292-4620-a13d-eb5c5205602a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1688.124719] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1688.124719] env[70013]: value = "task-4231799" [ 1688.124719] env[70013]: _type = "Task" [ 1688.124719] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1688.132564] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231799, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1688.172208] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]5221a0f5-c318-03d4-c699-23f8cf062cad, 'name': SearchDatastore_Task, 'duration_secs': 0.009899} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1688.172977] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb8a131a-83dc-433f-9c24-b5e1d4e53d23 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1688.178323] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1688.178323] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52265da0-2873-df95-346f-e4dbc4e85757" [ 1688.178323] env[70013]: _type = "Task" [ 1688.178323] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1688.186740] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52265da0-2873-df95-346f-e4dbc4e85757, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1688.634744] env[70013]: DEBUG oslo_vmware.api [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231799, 'name': PowerOnVM_Task, 'duration_secs': 0.414968} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1688.635016] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1688.635235] env[70013]: INFO nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Took 4.80 seconds to spawn the instance on the hypervisor. [ 1688.635422] env[70013]: DEBUG nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1688.636265] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1190fc26-dd37-4ed5-b626-cd1fb75bdc8c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1688.688545] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52265da0-2873-df95-346f-e4dbc4e85757, 'name': SearchDatastore_Task, 'duration_secs': 0.010391} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1688.688823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1688.689112] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1688.689377] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c8b68b9-e6aa-4d05-b866-f3d9e671772c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1688.695480] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1688.695480] env[70013]: value = "task-4231800" [ 1688.695480] env[70013]: _type = "Task" [ 1688.695480] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1688.703470] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231800, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1689.155246] env[70013]: INFO nova.compute.manager [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Took 9.57 seconds to build instance. [ 1689.206285] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231800, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472156} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1689.206551] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1689.206782] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1689.207090] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45ef60a1-ce05-461d-8ddd-66a3f9ec0a37 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1689.214011] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1689.214011] env[70013]: value = "task-4231801" [ 1689.214011] env[70013]: _type = "Task" [ 1689.214011] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1689.223665] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231801, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1689.658606] env[70013]: DEBUG oslo_concurrency.lockutils [None req-b0829d17-f0dc-4f98-8154-b6540719b594 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "538d6d33-5134-4574-9c28-31215aa6df60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.084s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1689.724355] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231801, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061176} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1689.724734] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1689.725562] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbb37fd-0f3d-4ae8-9e49-81d2b2cfc68e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1689.746215] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1689.746471] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0f1344c-dce9-46fa-bfd5-b4e7f6a59844 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1689.766153] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1689.766153] env[70013]: value = "task-4231802" [ 1689.766153] env[70013]: _type = "Task" [ 1689.766153] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1689.774308] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231802, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1690.276931] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231802, 'name': ReconfigVM_Task, 'duration_secs': 0.335536} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1690.277328] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Reconfigured VM instance instance-00000077 to attach disk [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1690.277835] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-615cfa78-4d3b-4765-a8fb-bd02d14fb620 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1690.285018] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1690.285018] env[70013]: value = "task-4231803" [ 1690.285018] env[70013]: _type = "Task" [ 1690.285018] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1690.293850] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231803, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1690.795990] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231803, 'name': Rename_Task, 'duration_secs': 0.136246} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1690.796352] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1690.796617] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d1429d4-9f03-41c9-9827-d406e174ddc8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1690.803178] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1690.803178] env[70013]: value = "task-4231804" [ 1690.803178] env[70013]: _type = "Task" [ 1690.803178] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1690.811402] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231804, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1691.313849] env[70013]: DEBUG oslo_vmware.api [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231804, 'name': PowerOnVM_Task, 'duration_secs': 0.475709} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1691.314322] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1691.314428] env[70013]: INFO nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Took 5.28 seconds to spawn the instance on the hypervisor. [ 1691.314558] env[70013]: DEBUG nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1691.315400] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4646e36a-e767-40f6-9d26-c98c07309708 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1691.831794] env[70013]: INFO nova.compute.manager [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Took 12.08 seconds to build instance. [ 1692.334528] env[70013]: DEBUG oslo_concurrency.lockutils [None req-6e98f80a-892f-41a5-9a4f-58b49b566634 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "f46c81c7-096d-4ac4-849c-608865ff6a39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.592s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1692.736503] env[70013]: INFO nova.compute.manager [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Rebuilding instance [ 1692.782235] env[70013]: DEBUG nova.compute.manager [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1692.783161] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10a112e-b4ba-4fac-8404-e0c9d47c770f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1693.799812] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1693.800282] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96b30bb4-3333-48e5-be09-60fb20c65c72 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1693.811201] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1693.811201] env[70013]: value = "task-4231805" [ 1693.811201] env[70013]: _type = "Task" [ 1693.811201] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1693.820559] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231805, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1694.321335] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231805, 'name': PowerOffVM_Task, 'duration_secs': 0.114947} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1694.321602] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1694.321835] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1694.322617] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecd58d4-b421-4c40-9d43-5e5c1eb241ba {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1694.329312] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1694.329526] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-597267ec-f93d-4e6e-bdc1-043e493aa063 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1694.357232] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1694.357466] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1694.357732] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleting the datastore file [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1694.358065] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a805d014-18ab-4bdd-a0a2-930456058f5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1694.365052] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1694.365052] env[70013]: value = "task-4231807" [ 1694.365052] env[70013]: _type = "Task" [ 1694.365052] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1694.373725] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231807, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1694.875465] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231807, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091187} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1694.875873] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1694.875930] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1694.876077] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1695.907634] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1695.907957] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1695.908085] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1695.908290] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1695.908445] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1695.908594] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1695.908810] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1695.908982] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1695.909188] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1695.909358] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1695.909533] env[70013]: DEBUG nova.virt.hardware [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1695.910415] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2d911b-7885-44e3-990b-b078e1c8ba3d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1695.918504] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cd9f8c-a341-4eae-8f3f-9833273f944a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1695.932393] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1695.938181] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1695.938747] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1695.938977] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-933f4157-46ca-4b54-9efc-fe58d0df52ca {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1695.956786] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1695.956786] env[70013]: value = "task-4231808" [ 1695.956786] env[70013]: _type = "Task" [ 1695.956786] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1695.965123] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231808, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1696.467698] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231808, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1696.967626] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231808, 'name': CreateVM_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1697.468835] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231808, 'name': CreateVM_Task, 'duration_secs': 1.270567} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1697.469043] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1697.469488] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1697.469649] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1697.469977] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1697.470333] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83951046-c94f-42db-a463-8d211d9279f1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1697.475227] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1697.475227] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]526693b1-e2f3-42bf-e267-550da9a2c637" [ 1697.475227] env[70013]: _type = "Task" [ 1697.475227] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1697.483019] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526693b1-e2f3-42bf-e267-550da9a2c637, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1697.985511] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]526693b1-e2f3-42bf-e267-550da9a2c637, 'name': SearchDatastore_Task, 'duration_secs': 0.01244} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1697.985936] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1697.986085] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1697.986336] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1697.986483] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1697.986663] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1697.986945] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd1ba399-67a3-40fc-9920-296279b34f39 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1697.996459] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1697.996633] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1697.997370] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d07edc5a-9f19-4dac-b875-b7373e37dd6a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1698.002703] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1698.002703] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52813f21-8b5d-6c77-c58b-a3663f5ffb88" [ 1698.002703] env[70013]: _type = "Task" [ 1698.002703] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1698.010209] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52813f21-8b5d-6c77-c58b-a3663f5ffb88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1698.513660] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52813f21-8b5d-6c77-c58b-a3663f5ffb88, 'name': SearchDatastore_Task, 'duration_secs': 0.008546} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1698.514459] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3d3c7f8-5596-47e8-a1b1-0e512fca84ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1698.519976] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1698.519976] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]529fda3b-15f1-7721-71ca-3ebb6af44d27" [ 1698.519976] env[70013]: _type = "Task" [ 1698.519976] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1698.527477] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529fda3b-15f1-7721-71ca-3ebb6af44d27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1699.031048] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]529fda3b-15f1-7721-71ca-3ebb6af44d27, 'name': SearchDatastore_Task, 'duration_secs': 0.012786} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1699.031440] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1699.031632] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1699.031813] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a826970-6692-4c7f-9366-33af2f80594a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1699.038818] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1699.038818] env[70013]: value = "task-4231809" [ 1699.038818] env[70013]: _type = "Task" [ 1699.038818] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1699.046231] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231809, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1699.548853] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231809, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468831} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1699.549193] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1699.549392] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1699.549642] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-457d99c9-92bd-4cd4-a87e-be38131f6b8b {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1699.556399] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1699.556399] env[70013]: value = "task-4231810" [ 1699.556399] env[70013]: _type = "Task" [ 1699.556399] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1699.564531] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231810, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1700.066233] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231810, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064626} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1700.066606] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1700.067232] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6caaf8-377e-4863-a520-ef20c66e2446 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1700.087158] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1700.087419] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6064cb0-395d-4b15-932a-10a6eae74486 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1700.106955] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1700.106955] env[70013]: value = "task-4231811" [ 1700.106955] env[70013]: _type = "Task" [ 1700.106955] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1700.114818] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231811, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1700.617050] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231811, 'name': ReconfigVM_Task, 'duration_secs': 0.26298} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1700.617351] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Reconfigured VM instance instance-00000077 to attach disk [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39/f46c81c7-096d-4ac4-849c-608865ff6a39.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1700.617962] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-855398d2-92e8-4e1b-a684-ca80dc91fb5e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1700.624451] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1700.624451] env[70013]: value = "task-4231812" [ 1700.624451] env[70013]: _type = "Task" [ 1700.624451] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1700.632280] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231812, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1701.136301] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231812, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1701.636224] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231812, 'name': Rename_Task} progress is 99%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1702.137422] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231812, 'name': Rename_Task, 'duration_secs': 1.123922} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1702.137812] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1702.137943] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1afaa865-4df6-416d-9f4b-6bd4aee768a7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1702.144653] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1702.144653] env[70013]: value = "task-4231813" [ 1702.144653] env[70013]: _type = "Task" [ 1702.144653] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1702.153776] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1702.654435] env[70013]: DEBUG oslo_vmware.api [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231813, 'name': PowerOnVM_Task, 'duration_secs': 0.453927} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1702.654810] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1702.655036] env[70013]: DEBUG nova.compute.manager [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1702.655831] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ed181a-f5b8-4f2a-b2be-f5040209ffde {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1703.172848] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1703.172848] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1703.172848] env[70013]: DEBUG nova.objects.instance [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1703.629568] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "f46c81c7-096d-4ac4-849c-608865ff6a39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1703.629842] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "f46c81c7-096d-4ac4-849c-608865ff6a39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1703.630086] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "f46c81c7-096d-4ac4-849c-608865ff6a39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1703.630309] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "f46c81c7-096d-4ac4-849c-608865ff6a39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1703.630485] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "f46c81c7-096d-4ac4-849c-608865ff6a39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1703.632686] env[70013]: INFO nova.compute.manager [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Terminating instance [ 1704.136868] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "refresh_cache-f46c81c7-096d-4ac4-849c-608865ff6a39" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1704.137093] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "refresh_cache-f46c81c7-096d-4ac4-849c-608865ff6a39" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1704.137269] env[70013]: DEBUG nova.network.neutron [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1704.179938] env[70013]: DEBUG oslo_concurrency.lockutils [None req-53aba21e-0ff9-4cee-97f3-a2ab4bcdaa4c tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1704.725666] env[70013]: DEBUG nova.network.neutron [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1704.781295] env[70013]: DEBUG nova.network.neutron [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1705.284284] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "refresh_cache-f46c81c7-096d-4ac4-849c-608865ff6a39" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1705.284746] env[70013]: DEBUG nova.compute.manager [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1705.285077] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1705.285865] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc634748-fc99-4619-9d92-6c051097d730 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1705.294163] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1705.294409] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f7541c9-ac8e-4cd5-a8c7-ca4c60fa37b4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1705.301169] env[70013]: DEBUG oslo_vmware.api [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1705.301169] env[70013]: value = "task-4231814" [ 1705.301169] env[70013]: _type = "Task" [ 1705.301169] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1705.309353] env[70013]: DEBUG oslo_vmware.api [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1705.813174] env[70013]: DEBUG oslo_vmware.api [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231814, 'name': PowerOffVM_Task, 'duration_secs': 0.178381} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1705.813881] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1705.813881] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1705.813881] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82c854fd-6b18-4ac8-9633-2fbdde277784 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1705.840933] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1705.841196] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1705.841361] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleting the datastore file [datastore2] f46c81c7-096d-4ac4-849c-608865ff6a39 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1705.841637] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d0d1be7a-c870-49f7-aebd-aff78466222c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1705.848770] env[70013]: DEBUG oslo_vmware.api [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1705.848770] env[70013]: value = "task-4231816" [ 1705.848770] env[70013]: _type = "Task" [ 1705.848770] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1705.856827] env[70013]: DEBUG oslo_vmware.api [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231816, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1706.360104] env[70013]: DEBUG oslo_vmware.api [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231816, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099609} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1706.360487] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1706.360614] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1706.360713] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1706.360890] env[70013]: INFO nova.compute.manager [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1706.361151] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1706.361347] env[70013]: DEBUG nova.compute.manager [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1706.361444] env[70013]: DEBUG nova.network.neutron [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1706.377177] env[70013]: DEBUG nova.network.neutron [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1706.880090] env[70013]: DEBUG nova.network.neutron [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1707.383022] env[70013]: INFO nova.compute.manager [-] [instance: f46c81c7-096d-4ac4-849c-608865ff6a39] Took 1.02 seconds to deallocate network for instance. [ 1707.890204] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1707.890501] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1707.890719] env[70013]: DEBUG nova.objects.instance [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lazy-loading 'resources' on Instance uuid f46c81c7-096d-4ac4-849c-608865ff6a39 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1708.514199] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ca6443-ab3b-4311-9289-d4e7101f8ddb {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1708.522204] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c08921b-8306-4678-b2b4-c4f4b7a2597c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1708.552541] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1331066b-4b3a-4508-a6e7-b4ae2c0445f7 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1708.560319] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d2727b-fb47-4ebf-a8c4-91f0f6259325 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1708.574167] env[70013]: DEBUG nova.compute.provider_tree [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1709.077698] env[70013]: DEBUG nova.scheduler.client.report [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1709.583387] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1709.601618] env[70013]: INFO nova.scheduler.client.report [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleted allocations for instance f46c81c7-096d-4ac4-849c-608865ff6a39 [ 1710.108746] env[70013]: DEBUG oslo_concurrency.lockutils [None req-85aa43a8-7fee-4044-9009-5e0ea34b9b9f tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "f46c81c7-096d-4ac4-849c-608865ff6a39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.479s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1710.367684] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "538d6d33-5134-4574-9c28-31215aa6df60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1710.367903] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "538d6d33-5134-4574-9c28-31215aa6df60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1710.368179] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "538d6d33-5134-4574-9c28-31215aa6df60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1710.368379] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "538d6d33-5134-4574-9c28-31215aa6df60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1710.368555] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "538d6d33-5134-4574-9c28-31215aa6df60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1710.370740] env[70013]: INFO nova.compute.manager [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Terminating instance [ 1710.874471] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "refresh_cache-538d6d33-5134-4574-9c28-31215aa6df60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1710.874851] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquired lock "refresh_cache-538d6d33-5134-4574-9c28-31215aa6df60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1710.874851] env[70013]: DEBUG nova.network.neutron [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1711.392890] env[70013]: DEBUG nova.network.neutron [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1711.434644] env[70013]: DEBUG nova.network.neutron [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1711.936829] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Releasing lock "refresh_cache-538d6d33-5134-4574-9c28-31215aa6df60" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1711.937313] env[70013]: DEBUG nova.compute.manager [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1711.937570] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1711.938559] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62dbd47e-d164-4854-b5f2-38200dc1a063 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1711.946431] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1711.946668] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-047e64fa-3db9-4f17-a68a-9c72ff4c4a07 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1711.965760] env[70013]: DEBUG oslo_vmware.api [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1711.965760] env[70013]: value = "task-4231817" [ 1711.965760] env[70013]: _type = "Task" [ 1711.965760] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1711.976320] env[70013]: DEBUG oslo_vmware.api [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1712.476231] env[70013]: DEBUG oslo_vmware.api [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231817, 'name': PowerOffVM_Task, 'duration_secs': 0.12023} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1712.476487] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1712.476655] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1712.476898] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-edcaa457-1f31-4e5e-8b19-8db744941aa4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1712.502303] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1712.502534] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1712.502717] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleting the datastore file [datastore2] 538d6d33-5134-4574-9c28-31215aa6df60 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1712.502984] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e58ee69-4566-4cb4-acd4-b7c7ba4f1a57 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1712.509708] env[70013]: DEBUG oslo_vmware.api [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for the task: (returnval){ [ 1712.509708] env[70013]: value = "task-4231819" [ 1712.509708] env[70013]: _type = "Task" [ 1712.509708] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1712.517525] env[70013]: DEBUG oslo_vmware.api [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1713.019797] env[70013]: DEBUG oslo_vmware.api [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Task: {'id': task-4231819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172028} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1713.020217] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1713.020267] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1713.020411] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1713.020585] env[70013]: INFO nova.compute.manager [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1713.020823] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1713.021020] env[70013]: DEBUG nova.compute.manager [-] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1713.021119] env[70013]: DEBUG nova.network.neutron [-] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1713.036382] env[70013]: DEBUG nova.network.neutron [-] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1713.538841] env[70013]: DEBUG nova.network.neutron [-] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1714.041714] env[70013]: INFO nova.compute.manager [-] [instance: 538d6d33-5134-4574-9c28-31215aa6df60] Took 1.02 seconds to deallocate network for instance. [ 1714.548586] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1714.548879] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1714.549123] env[70013]: DEBUG nova.objects.instance [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lazy-loading 'resources' on Instance uuid 538d6d33-5134-4574-9c28-31215aa6df60 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1715.156917] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c0c866-3a6f-4270-9e5f-30c64c2f17df {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1715.164838] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e654b56-ad4f-4a8d-a94e-4c2c8e9c9b56 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1715.195353] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c625ac21-3a96-471d-9a01-498dc7593435 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1715.202796] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b5be41-dc14-4584-a762-4b99eef5ff02 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1715.216511] env[70013]: DEBUG nova.compute.provider_tree [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1715.719907] env[70013]: DEBUG nova.scheduler.client.report [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1716.225733] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1716.246738] env[70013]: INFO nova.scheduler.client.report [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Deleted allocations for instance 538d6d33-5134-4574-9c28-31215aa6df60 [ 1716.756392] env[70013]: DEBUG oslo_concurrency.lockutils [None req-091953a6-c49e-4898-a2b1-11f1538a4893 tempest-ServerShowV247Test-175835694 tempest-ServerShowV247Test-175835694-project-member] Lock "538d6d33-5134-4574-9c28-31215aa6df60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.388s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1718.942296] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "eb3c8828-b913-44ad-8599-5ae35e059126" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1718.942632] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "eb3c8828-b913-44ad-8599-5ae35e059126" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1719.445329] env[70013]: DEBUG nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Starting instance... {{(pid=70013) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1719.964319] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1719.964591] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1719.966116] env[70013]: INFO nova.compute.claims [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1721.091546] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40555e9-0124-4a89-acf6-422a294a852e {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1721.099130] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d484e0-bb86-4586-ac0c-f634c3987417 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1721.129668] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318fd547-918f-4dca-863b-fe1a98bf25c4 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1721.136942] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9151c74f-c707-4406-a97f-5b302e53b427 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1721.150204] env[70013]: DEBUG nova.compute.provider_tree [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1721.653948] env[70013]: DEBUG nova.scheduler.client.report [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1722.158893] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1722.159508] env[70013]: DEBUG nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Start building networks asynchronously for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1722.664589] env[70013]: DEBUG nova.compute.utils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Using /dev/sd instead of None {{(pid=70013) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1722.666105] env[70013]: DEBUG nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Not allocating networking since 'none' was specified. {{(pid=70013) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1723.167878] env[70013]: DEBUG nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Start building block device mappings for instance. {{(pid=70013) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1724.178092] env[70013]: DEBUG nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Start spawning the instance on the hypervisor. {{(pid=70013) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1724.204849] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1724.205112] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1724.205273] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1724.205460] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1724.205607] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1724.205835] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1724.205961] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1724.206161] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1724.206358] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1724.206532] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1724.206709] env[70013]: DEBUG nova.virt.hardware [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1724.207630] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadeb29c-5229-40cf-a5c3-67c3e63ab0b8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1724.215954] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d43cdc-a964-49c5-8885-a27d0716461c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1724.230879] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1724.236330] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Creating folder: Project (5aed9b0c694b4ab2b4adea950e3bcdf6). Parent ref: group-v836999. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1724.236609] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2842b6ad-27e3-439a-abf1-56b71efa11f3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1724.249018] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Created folder: Project (5aed9b0c694b4ab2b4adea950e3bcdf6) in parent group-v836999. [ 1724.249239] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Creating folder: Instances. Parent ref: group-v837267. {{(pid=70013) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1724.249492] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55a37ba2-f4ca-4094-8131-633c0e95582d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1724.260087] env[70013]: INFO nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Created folder: Instances in parent group-v837267. [ 1724.260347] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1724.260553] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1724.260753] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43fb17f0-b202-427a-8b8f-a177ad96df17 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1724.278568] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1724.278568] env[70013]: value = "task-4231822" [ 1724.278568] env[70013]: _type = "Task" [ 1724.278568] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1724.286449] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231822, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1724.789031] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231822, 'name': CreateVM_Task, 'duration_secs': 0.270858} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1724.789031] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1724.789031] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1724.789288] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1724.789511] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1724.789771] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44132e76-0c09-415f-acb0-0a50a7bbf292 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1724.795138] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1724.795138] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3b394-14aa-f0a7-b832-7158b96286da" [ 1724.795138] env[70013]: _type = "Task" [ 1724.795138] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1724.803121] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3b394-14aa-f0a7-b832-7158b96286da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1725.305778] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52d3b394-14aa-f0a7-b832-7158b96286da, 'name': SearchDatastore_Task, 'duration_secs': 0.009666} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1725.306170] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1725.306335] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1725.306573] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1725.306721] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1725.306898] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1725.307185] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3509f265-bf0b-41d7-b1f4-08dc3bdf81a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1725.316215] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1725.316404] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1725.317153] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-893be667-5b9c-42d9-be97-0c2950ac19b3 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1725.322885] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1725.322885] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b2f5b-9888-5798-96a0-7ea398934eda" [ 1725.322885] env[70013]: _type = "Task" [ 1725.322885] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1725.330841] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b2f5b-9888-5798-96a0-7ea398934eda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1725.833926] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]525b2f5b-9888-5798-96a0-7ea398934eda, 'name': SearchDatastore_Task, 'duration_secs': 0.009655} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1725.834705] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82279154-096b-482d-92dc-cfd88ddf4d9f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1725.840542] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1725.840542] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52398e73-9a79-6fc5-017b-b1dba60d73a1" [ 1725.840542] env[70013]: _type = "Task" [ 1725.840542] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1725.849216] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52398e73-9a79-6fc5-017b-b1dba60d73a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1726.339591] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1726.351757] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52398e73-9a79-6fc5-017b-b1dba60d73a1, 'name': SearchDatastore_Task, 'duration_secs': 0.010829} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1726.352844] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1726.353141] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1726.353411] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f29295c-eaae-4904-a477-f5b806e70571 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1726.360313] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1726.360313] env[70013]: value = "task-4231823" [ 1726.360313] env[70013]: _type = "Task" [ 1726.360313] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1726.368750] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1726.842938] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1726.843230] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1726.843411] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1726.843585] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=70013) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1726.844488] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c3a559-0271-4e7f-acb4-a85a03ef0828 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1726.852986] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612f133b-0e9d-4233-af38-0db65edad36d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1726.869663] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048e7e86-38cd-4f24-844d-62b38f3ca180 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1726.879299] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231823, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451142} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1726.880274] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88762fa-b8ab-4eda-baa7-df104e9041ab {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1726.883279] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1726.883503] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1726.883734] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-898106f5-978e-4cf0-b88f-c37f3066ba24 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1726.891363] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1726.891363] env[70013]: value = "task-4231824" [ 1726.891363] env[70013]: _type = "Task" [ 1726.891363] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1726.916966] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180478MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=70013) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1726.917151] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1726.917389] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1726.928141] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231824, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1727.435824] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231824, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060749} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1727.436640] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1727.437880] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa157270-2a06-4614-b9f2-4232315e9c85 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1727.471974] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1727.472737] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4af76ef-13ab-47ba-b2a6-74f36e7f4d2c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1727.507339] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1727.507339] env[70013]: value = "task-4231825" [ 1727.507339] env[70013]: _type = "Task" [ 1727.507339] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1727.520111] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231825, 'name': ReconfigVM_Task} progress is 6%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1727.957292] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance f12b8575-5082-4be9-9bf5-f4279860d19d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.957575] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 789fffd9-6725-4bf6-9144-dd603b0a521f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.957813] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance ef64a05a-b514-4c35-81d3-664ae1ad3ff1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.958043] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9137a6bc-8f1e-47b7-b1ec-5b4035d36fff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.958324] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 2aefaa91-c439-486a-8b19-c6f45f52583f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.958529] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 39557c50-866a-4510-b840-b1a6a3e3890e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.958753] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance c6e9e2dc-2cbb-4be7-aa73-fcb30dfa2dc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.958974] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance 9a61dedd-3764-4bd9-a300-480cc7d14a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.959228] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Instance eb3c8828-b913-44ad-8599-5ae35e059126 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=70013) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1737}} [ 1727.959564] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1727.959829] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=70013) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1728.020907] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231825, 'name': ReconfigVM_Task, 'duration_secs': 0.309832} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1728.023625] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Reconfigured VM instance instance-00000078 to attach disk [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1728.025035] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2056bc0b-d41b-43ff-9521-3b33355cc63d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1728.032991] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1728.032991] env[70013]: value = "task-4231826" [ 1728.032991] env[70013]: _type = "Task" [ 1728.032991] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1728.044912] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231826, 'name': Rename_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1728.096496] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ef739a-a004-4c4c-85b2-d53b70aaa728 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1728.104285] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab11fec-d85c-4196-beb0-16869b9e175d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1728.137093] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26d264b-f126-4c48-81ef-93387b26ff0f {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1728.145619] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfaaf0e-47de-49aa-a189-169d6ac5fa41 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1728.159918] env[70013]: DEBUG nova.compute.provider_tree [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1728.544257] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231826, 'name': Rename_Task, 'duration_secs': 0.149153} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1728.544685] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1728.544753] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f6314f7-d21e-4fe5-9856-bd30598f00e0 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1728.551706] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1728.551706] env[70013]: value = "task-4231827" [ 1728.551706] env[70013]: _type = "Task" [ 1728.551706] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1728.559647] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231827, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1728.662880] env[70013]: DEBUG nova.scheduler.client.report [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1729.061806] env[70013]: DEBUG oslo_vmware.api [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231827, 'name': PowerOnVM_Task, 'duration_secs': 0.406209} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1729.061956] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1729.062165] env[70013]: INFO nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Took 4.88 seconds to spawn the instance on the hypervisor. [ 1729.062348] env[70013]: DEBUG nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1729.063139] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa8c3f3-a926-48d5-82d2-3c014b9a2b7c {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1729.168080] env[70013]: DEBUG nova.compute.resource_tracker [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=70013) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1729.168294] env[70013]: DEBUG oslo_concurrency.lockutils [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.251s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1729.581339] env[70013]: INFO nova.compute.manager [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Took 9.63 seconds to build instance. [ 1730.084226] env[70013]: DEBUG oslo_concurrency.lockutils [None req-dde71fde-9180-4189-9a6f-539b4d28b723 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "eb3c8828-b913-44ad-8599-5ae35e059126" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.141s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1731.073096] env[70013]: INFO nova.compute.manager [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Rebuilding instance [ 1731.113533] env[70013]: DEBUG nova.compute.manager [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1731.114446] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7615ac-b1be-4ba1-aebe-9ae47c5ea394 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1731.169313] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1731.169527] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1731.169672] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1731.169821] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1731.170019] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1731.170175] env[70013]: DEBUG nova.compute.manager [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=70013) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11183}} [ 1731.334866] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1732.128506] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1732.128912] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2536409-267d-4a87-8057-8871f6bd6aef {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1732.136653] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1732.136653] env[70013]: value = "task-4231828" [ 1732.136653] env[70013]: _type = "Task" [ 1732.136653] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1732.145154] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1732.646278] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231828, 'name': PowerOffVM_Task, 'duration_secs': 0.126166} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1732.646548] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1732.647266] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1732.648058] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f38557-260b-46de-a02a-c2fa0c28caa8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1732.654756] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1732.654985] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11f7c4ab-26f7-4365-8c16-5ca90c216b94 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1732.685730] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1732.685978] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1732.686153] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Deleting the datastore file [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1732.686457] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a4fa6fc6-928c-491b-8df8-f315ded3ef55 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1732.692389] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1732.692389] env[70013]: value = "task-4231830" [ 1732.692389] env[70013]: _type = "Task" [ 1732.692389] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1732.700345] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1733.202480] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098814} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1733.202865] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1733.202981] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1733.203182] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1734.233519] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-19T11:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-19T11:30:53Z,direct_url=,disk_format='vmdk',id=0ddeef65-a14c-4065-97bb-58b607968d19,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1e6c9228801846fd94db2231de2054c8',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-19T11:30:53Z,virtual_size=,visibility=), allow threads: False {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1734.233829] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Flavor limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1734.233974] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Image limits 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1734.234181] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Flavor pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1734.234334] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Image pref 0:0:0 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1734.234487] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=70013) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1734.234692] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1734.234893] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1734.235089] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Got 1 possible topologies {{(pid=70013) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1734.235261] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1734.235437] env[70013]: DEBUG nova.virt.hardware [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=70013) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1734.236350] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91aee96-46a3-42bf-be2c-ac74f7417425 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1734.244686] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebaa25bd-cdf0-495b-b2d5-9f904dd8d388 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1734.259634] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Instance VIF info [] {{(pid=70013) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1734.265165] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1734.265383] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Creating VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1734.265588] env[70013]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0de229eb-495a-4915-854c-1cb3cdf2a7c6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1734.283423] env[70013]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1734.283423] env[70013]: value = "task-4231831" [ 1734.283423] env[70013]: _type = "Task" [ 1734.283423] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1734.291706] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231831, 'name': CreateVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1734.333520] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1734.793017] env[70013]: DEBUG oslo_vmware.api [-] Task: {'id': task-4231831, 'name': CreateVM_Task, 'duration_secs': 0.498738} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1734.793226] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Created VM on the ESX host {{(pid=70013) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1734.793620] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1734.793787] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1734.794148] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1734.794418] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02dcc631-31f6-4dd8-8250-7b163daff3a1 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1734.799175] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1734.799175] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab571f-5d05-1927-c3b0-a56d3ebd22aa" [ 1734.799175] env[70013]: _type = "Task" [ 1734.799175] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1734.806841] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab571f-5d05-1927-c3b0-a56d3ebd22aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1735.309982] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52ab571f-5d05-1927-c3b0-a56d3ebd22aa, 'name': SearchDatastore_Task, 'duration_secs': 0.009525} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1735.310564] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1735.310775] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Processing image 0ddeef65-a14c-4065-97bb-58b607968d19 {{(pid=70013) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1735.311118] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1735.311358] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1735.311622] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1735.311946] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-52246c25-05e6-4aea-890b-640075abf5e6 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1735.320267] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=70013) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1735.320525] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=70013) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1735.321309] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70f09faa-89e3-41cb-91d2-04fe78e7285d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1735.326289] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1735.326289] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3a405-2f64-0d40-6124-189173d2a975" [ 1735.326289] env[70013]: _type = "Task" [ 1735.326289] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1735.333753] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3a405-2f64-0d40-6124-189173d2a975, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1735.837802] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52f3a405-2f64-0d40-6124-189173d2a975, 'name': SearchDatastore_Task, 'duration_secs': 0.008451} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1735.838667] env[70013]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29fe61d5-f0c6-4293-99c2-1ac8c735cc13 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1735.844598] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1735.844598] env[70013]: value = "session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1dc22-1022-5ff4-46c7-14a9b305d49d" [ 1735.844598] env[70013]: _type = "Task" [ 1735.844598] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1735.852531] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1dc22-1022-5ff4-46c7-14a9b305d49d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1736.339783] env[70013]: DEBUG oslo_service.periodic_task [None req-0d43694d-6a2d-417b-a1ff-67024b771fe7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=70013) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1736.354394] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': session[52df8453-14cf-279c-2a8e-fe5645d8684b]52e1dc22-1022-5ff4-46c7-14a9b305d49d, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1736.354658] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1736.354911] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1736.355178] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15c2e1dd-9c25-426b-9a41-57c2d6ac2b81 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1736.362163] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1736.362163] env[70013]: value = "task-4231832" [ 1736.362163] env[70013]: _type = "Task" [ 1736.362163] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1736.369652] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231832, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1736.873213] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231832, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432259} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1736.873457] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0ddeef65-a14c-4065-97bb-58b607968d19/0ddeef65-a14c-4065-97bb-58b607968d19.vmdk to [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk {{(pid=70013) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1736.873682] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Extending root virtual disk to 1048576 {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1736.873946] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bed33dfc-50bb-4b3d-a0d9-2d1e0731a557 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1736.880678] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1736.880678] env[70013]: value = "task-4231833" [ 1736.880678] env[70013]: _type = "Task" [ 1736.880678] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1736.890189] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231833, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1737.390647] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231833, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067695} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1737.391105] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Extended root virtual disk {{(pid=70013) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1737.391699] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfdf570-6ced-46dd-a3b7-0ee97833d45d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1737.412211] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1737.412451] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccaf9d08-56da-4b88-a59c-9b67f1f38ccd {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1737.432676] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1737.432676] env[70013]: value = "task-4231834" [ 1737.432676] env[70013]: _type = "Task" [ 1737.432676] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1737.441355] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231834, 'name': ReconfigVM_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1737.943162] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231834, 'name': ReconfigVM_Task, 'duration_secs': 0.304962} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1737.943500] env[70013]: DEBUG nova.virt.vmwareapi.volumeops [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Reconfigured VM instance instance-00000078 to attach disk [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126/eb3c8828-b913-44ad-8599-5ae35e059126.vmdk or device None with type sparse {{(pid=70013) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1737.944175] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-344f67af-ed15-4cbb-89b5-9d907f4b848d {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1737.951974] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1737.951974] env[70013]: value = "task-4231835" [ 1737.951974] env[70013]: _type = "Task" [ 1737.951974] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1737.961140] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231835, 'name': Rename_Task} progress is 5%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1738.462813] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231835, 'name': Rename_Task, 'duration_secs': 0.13614} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1738.463212] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powering on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1738.463337] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e885672f-3d49-4d04-82be-3cc7ff162a50 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1738.470477] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1738.470477] env[70013]: value = "task-4231836" [ 1738.470477] env[70013]: _type = "Task" [ 1738.470477] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1738.478420] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231836, 'name': PowerOnVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1738.980092] env[70013]: DEBUG oslo_vmware.api [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231836, 'name': PowerOnVM_Task, 'duration_secs': 0.407436} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1738.980360] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powered on the VM {{(pid=70013) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1738.980576] env[70013]: DEBUG nova.compute.manager [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Checking state {{(pid=70013) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1738.981382] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01f85c9-725f-46cc-aae4-388e06004ed2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1739.498823] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1739.499114] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1739.499297] env[70013]: DEBUG nova.objects.instance [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Trying to apply a migration context that does not seem to be set for this instance {{(pid=70013) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1739.782066] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "eb3c8828-b913-44ad-8599-5ae35e059126" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1739.782325] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "eb3c8828-b913-44ad-8599-5ae35e059126" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1739.782543] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "eb3c8828-b913-44ad-8599-5ae35e059126-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1739.782733] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "eb3c8828-b913-44ad-8599-5ae35e059126-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1739.782903] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "eb3c8828-b913-44ad-8599-5ae35e059126-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1739.785115] env[70013]: INFO nova.compute.manager [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Terminating instance [ 1740.288755] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "refresh_cache-eb3c8828-b913-44ad-8599-5ae35e059126" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1740.288970] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquired lock "refresh_cache-eb3c8828-b913-44ad-8599-5ae35e059126" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1740.289140] env[70013]: DEBUG nova.network.neutron [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Building network info cache for instance {{(pid=70013) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1740.507780] env[70013]: DEBUG oslo_concurrency.lockutils [None req-d738306f-01e4-46cf-98b8-48101e47a984 tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1740.808752] env[70013]: DEBUG nova.network.neutron [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1740.868013] env[70013]: DEBUG nova.network.neutron [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1741.370593] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Releasing lock "refresh_cache-eb3c8828-b913-44ad-8599-5ae35e059126" {{(pid=70013) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1741.371117] env[70013]: DEBUG nova.compute.manager [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Start destroying the instance on the hypervisor. {{(pid=70013) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1741.371365] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Destroying instance {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1741.372482] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5f8bae-ea7a-46eb-a2d0-282a159e85a9 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1741.380867] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powering off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1741.381192] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74a79513-22a3-4bdf-bce3-1e6bfa27b141 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1741.388055] env[70013]: DEBUG oslo_vmware.api [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1741.388055] env[70013]: value = "task-4231837" [ 1741.388055] env[70013]: _type = "Task" [ 1741.388055] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1741.397358] env[70013]: DEBUG oslo_vmware.api [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231837, 'name': PowerOffVM_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1741.898515] env[70013]: DEBUG oslo_vmware.api [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231837, 'name': PowerOffVM_Task, 'duration_secs': 0.185826} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1741.898915] env[70013]: DEBUG nova.virt.vmwareapi.vm_util [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Powered off the VM {{(pid=70013) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1741.898962] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Unregistering the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1741.899211] env[70013]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4646e93-2432-40d6-b970-c0a37a7da042 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1741.925927] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Unregistered the VM {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1741.926162] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Deleting contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1741.926400] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Deleting the datastore file [datastore2] eb3c8828-b913-44ad-8599-5ae35e059126 {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1741.926681] env[70013]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5adf956-bed8-4db9-a1f9-bf66227718f5 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1741.933813] env[70013]: DEBUG oslo_vmware.api [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for the task: (returnval){ [ 1741.933813] env[70013]: value = "task-4231839" [ 1741.933813] env[70013]: _type = "Task" [ 1741.933813] env[70013]: } to complete. {{(pid=70013) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1741.941869] env[70013]: DEBUG oslo_vmware.api [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1742.443358] env[70013]: DEBUG oslo_vmware.api [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Task: {'id': task-4231839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11866} completed successfully. {{(pid=70013) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1742.443620] env[70013]: DEBUG nova.virt.vmwareapi.ds_util [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Deleted the datastore file {{(pid=70013) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1742.443806] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Deleted contents of the VM from datastore datastore2 {{(pid=70013) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1742.443984] env[70013]: DEBUG nova.virt.vmwareapi.vmops [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Instance destroyed {{(pid=70013) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1742.444179] env[70013]: INFO nova.compute.manager [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1742.444419] env[70013]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=70013) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1742.444613] env[70013]: DEBUG nova.compute.manager [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Deallocating network for instance {{(pid=70013) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1742.444705] env[70013]: DEBUG nova.network.neutron [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] deallocate_for_instance() {{(pid=70013) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1742.460084] env[70013]: DEBUG nova.network.neutron [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Instance cache missing network info. {{(pid=70013) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1742.963062] env[70013]: DEBUG nova.network.neutron [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Updating instance_info_cache with network_info: [] {{(pid=70013) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1743.464935] env[70013]: INFO nova.compute.manager [-] [instance: eb3c8828-b913-44ad-8599-5ae35e059126] Took 1.02 seconds to deallocate network for instance. [ 1743.971949] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1743.972405] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1743.972483] env[70013]: DEBUG nova.objects.instance [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lazy-loading 'resources' on Instance uuid eb3c8828-b913-44ad-8599-5ae35e059126 {{(pid=70013) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1744.592777] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bb7971-31e2-42ee-8be0-4b8a9233cd3a {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1744.601063] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae8f99c-6e04-40c6-b36a-f42544eb3ac2 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1744.632671] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e9c9a0-0239-4dcb-9a0e-554d989e2549 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1744.640481] env[70013]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6ef190-e4c7-4f8f-8d64-26fc7fbf7db8 {{(pid=70013) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1744.654267] env[70013]: DEBUG nova.compute.provider_tree [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Inventory has not changed in ProviderTree for provider: dc0a4c9d-b611-453e-a900-41e280c29b95 {{(pid=70013) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1745.157231] env[70013]: DEBUG nova.scheduler.client.report [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Inventory has not changed for provider dc0a4c9d-b611-453e-a900-41e280c29b95 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=70013) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1745.661843] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1745.682201] env[70013]: INFO nova.scheduler.client.report [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Deleted allocations for instance eb3c8828-b913-44ad-8599-5ae35e059126 [ 1746.190992] env[70013]: DEBUG oslo_concurrency.lockutils [None req-25b9d2ea-be01-43c3-9d65-f775fae88c1c tempest-ServersListShow298Test-1464588438 tempest-ServersListShow298Test-1464588438-project-member] Lock "eb3c8828-b913-44ad-8599-5ae35e059126" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.409s {{(pid=70013) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}}